July 26, 2019

2299 words 11 mins read

Paper Group NANR 18

Paper Group NANR 18

SHAKKIL: An Automatic Diacritization System for Modern Standard Arabic Texts. Integrating Order Information and Event Relation for Script Event Prediction. Using NLP for Enhancing Second Language Acquisition. Rational Distortions of Learners’ Linguistic Input. A data-driven model of explanations for a chatbot that helps to practice conversation in …

SHAKKIL: An Automatic Diacritization System for Modern Standard Arabic Texts

Title SHAKKIL: An Automatic Diacritization System for Modern Standard Arabic Texts
Authors Amany Fashwan, Sameh Alansary
Abstract This paper sheds light on a system that would be able to diacritize Arabic texts automatically (SHAKKIL). In this system, the diacritization problem will be handled through two levels; morphological and syntactic processing levels. The adopted morphological disambiguation algorithm depends on four layers; Uni-morphological form layer, rule-based morphological disambiguation layer, statistical-based disambiguation layer and Out Of Vocabulary (OOV) layer. The adopted syntactic disambiguation algorithms is concerned with detecting the case ending diacritics depending on a rule based approach simulating the shallow parsing technique. This will be achieved using an annotated corpus for extracting the Arabic linguistic rules, building the language models and testing the system output. This system is considered as a good trial of the interaction between rule-based approach and statistical approach, where the rules can help the statistics in detecting the right diacritization and vice versa. At this point, the morphological Word Error Rate (WER) is 4.56{%} while the morphological Diacritic Error Rate (DER) is 1.88{%} and the syntactic WER is 9.36{%}. The best WER is 14.78{%} compared to the best-published results, of (Abandah, 2015); 11.68{%}, (Rashwan, et al., 2015); 12.90{%} and (Metwally, Rashwan, {&} Atiya, 2016); 13.70{%}.
Tasks Information Retrieval, Machine Translation
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-1311/
PDF https://www.aclweb.org/anthology/W17-1311
PWC https://paperswithcode.com/paper/shakkil-an-automatic-diacritization-system
Repo
Framework

Integrating Order Information and Event Relation for Script Event Prediction

Title Integrating Order Information and Event Relation for Script Event Prediction
Authors Zhongqing Wang, Yue Zhang, Ching-Yun Chang
Abstract There has been a recent line of work automatically learning scripts from unstructured texts, by modeling narrative event chains. While the dominant approach group events using event pair relations, LSTMs have been used to encode full chains of narrative events. The latter has the advantage of learning long-range temporal orders, yet the former is more adaptive to partial orders. We propose a neural model that leverages the advantages of both methods, by using LSTM hidden states as features for event pair modelling. A dynamic memory network is utilized to automatically induce weights on existing events for inferring a subsequent event. Standard evaluation shows that our method significantly outperforms both methods above, giving the best results reported so far.
Tasks Language Modelling
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1006/
PDF https://www.aclweb.org/anthology/D17-1006
PWC https://paperswithcode.com/paper/integrating-order-information-and-event
Repo
Framework

Using NLP for Enhancing Second Language Acquisition

Title Using NLP for Enhancing Second Language Acquisition
Authors Leonardo Zilio, Rodrigo Wilkens, C{'e}drick Fairon
Abstract This study presents SMILLE, a system that draws on the Noticing Hypothesis and on input enhancements, addressing the lack of salience of grammatical infor mation in online documents chosen by a given user. By means of input enhancements, the system can draw the user{'}s attention to grammar, which could possibly lead to a higher intake per input ratio for metalinguistic information. The system receives as input an online document and submits it to a combined processing of parser and hand-written rules for detecting its grammatical structures. The input text can be freely chosen by the user, providing a more engaging experience and reflecting the user{'}s interests. The system can enhance a total of 107 fine-grained types of grammatical structures that are based on the CEFR. An evaluation of some of those structures resulted in an overall precision of 87{%}.
Tasks Language Acquisition
Published 2017-09-01
URL https://www.aclweb.org/anthology/R17-1107/
PDF https://doi.org/10.26615/978-954-452-049-6_107
PWC https://paperswithcode.com/paper/using-nlp-for-enhancing-second-language
Repo
Framework

Rational Distortions of Learners’ Linguistic Input

Title Rational Distortions of Learners’ Linguistic Input
Authors Naomi Feldman
Abstract Language acquisition can be modeled as a statistical inference problem: children use sentences and sounds in their input to infer linguistic structure. However, in many cases, children learn from data whose statistical structure is distorted relative to the language they are learning. Such distortions can arise either in the input itself, or as a result of children{'}s immature strategies for encoding their input. This work examines several cases in which the statistical structure of children{'}s input differs from the language being learned. Analyses show that these distortions of the input can be accounted for with a statistical learning framework by carefully considering the inference problems that learners solve during language acquisition
Tasks Language Acquisition
Published 2017-08-01
URL https://www.aclweb.org/anthology/K17-1002/
PDF https://www.aclweb.org/anthology/K17-1002
PWC https://paperswithcode.com/paper/rational-distortions-of-learners-linguistic
Repo
Framework

A data-driven model of explanations for a chatbot that helps to practice conversation in a foreign language

Title A data-driven model of explanations for a chatbot that helps to practice conversation in a foreign language
Authors Sviatlana H{"o}hn
Abstract This article describes a model of other-initiated self-repair for a chatbot that helps to practice conversation in a foreign language. The model was developed using a corpus of instant messaging conversations between German native and non-native speakers. Conversation Analysis helped to create computational models from a small number of examples. The model has been validated in an AIML-based chatbot. Unlike typical retrieval-based dialogue systems, the explanations are generated at run-time from a linguistic database.
Tasks Chatbot, Language Acquisition
Published 2017-08-01
URL https://www.aclweb.org/anthology/W17-5547/
PDF https://www.aclweb.org/anthology/W17-5547
PWC https://paperswithcode.com/paper/a-data-driven-model-of-explanations-for-a
Repo
Framework

Neoveille, a Web Platform for Neologism Tracking

Title Neoveille, a Web Platform for Neologism Tracking
Authors Emmanuel Cartier
Abstract This paper details a software designed to track neologisms in seven languages through newspapers monitor corpora. The platform combines state-of-the-art processes to track linguistic changes and a web platform for linguists to create and manage their corpora, accept or reject automatically identified neologisms, describe linguistically the accepted neologisms and follow their lifecycle on the monitor corpora. In the following, after a short state-of-the-art in Neologism Retrieval, Analysis and Life-tracking, we describe the overall architecture of the system. The platform can be freely browsed at \url{www.neoveille.org} where detailed presentation is given. Access tothe editing modules is available upon request.
Tasks
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-3024/
PDF https://www.aclweb.org/anthology/E17-3024
PWC https://paperswithcode.com/paper/neoveille-a-web-platform-for-neologism
Repo
Framework

Learning Deep Architectures via Generalized Whitened Neural Networks

Title Learning Deep Architectures via Generalized Whitened Neural Networks
Authors Ping Luo
Abstract Whitened Neural Network (WNN) is a recent advanced deep architecture, which improves convergence and generalization of canonical neural networks by whitening their internal hidden representation. However, the whitening transformation increases computation time. Unlike WNN that reduced runtime by performing whitening every thousand iterations, which degenerates convergence due to the ill conditioning, we present generalized WNN (GWNN), which has three appealing properties. First, GWNN is able to learn compact representation to reduce computations. Second, it enables whitening transformation to be performed in a short period, preserving good conditioning. Third, we propose a data-independent estimation of the covariance matrix to further improve computational efficiency. Extensive experiments on various datasets demonstrate the benefits of GWNN.
Tasks
Published 2017-08-01
URL https://icml.cc/Conferences/2017/Schedule?showEvent=488
PDF http://proceedings.mlr.press/v70/luo17a/luo17a.pdf
PWC https://paperswithcode.com/paper/learning-deep-architectures-via-generalized
Repo
Framework

End to End Dialog System for Telugu

Title End to End Dialog System for Telugu
Authors D, Prathyusha a, Prathyusha Jwalapuram, Manish Shrivastava
Abstract
Tasks
Published 2017-12-01
URL https://www.aclweb.org/anthology/W17-7533/
PDF https://www.aclweb.org/anthology/W17-7533
PWC https://paperswithcode.com/paper/end-to-end-dialog-system-for-telugu
Repo
Framework

Language Modeling with Recurrent Highway Hypernetworks

Title Language Modeling with Recurrent Highway Hypernetworks
Authors Joseph Suarez
Abstract We present extensive experimental and theoretical support for the efficacy of recurrent highway networks (RHNs) and recurrent hypernetworks complimentary to the original works. Where the original RHN work primarily provides theoretical treatment of the subject, we demonstrate experimentally that RHNs benefit from far better gradient flow than LSTMs in addition to their improved task accuracy. The original hypernetworks work presents detailed experimental results but leaves several theoretical issues unresolved–we consider these in depth and frame several feasible solutions that we believe will yield further gains in the future. We demonstrate that these approaches are complementary: by combining RHNs and hypernetworks, we make a significant improvement over current state-of-the-art character-level language modeling performance on Penn Treebank while relying on much simpler regularization. Finally, we argue for RHNs as a drop-in replacement for LSTMs (analogous to LSTMs for vanilla RNNs) and for hypernetworks as a de-facto augmentation (analogous to attention) for recurrent architectures.
Tasks Language Modelling
Published 2017-12-01
URL http://papers.nips.cc/paper/6919-language-modeling-with-recurrent-highway-hypernetworks
PDF http://papers.nips.cc/paper/6919-language-modeling-with-recurrent-highway-hypernetworks.pdf
PWC https://paperswithcode.com/paper/language-modeling-with-recurrent-highway
Repo
Framework

High-Precision Automated Reconstruction of Neurons with Flood-filling Networks

Title High-Precision Automated Reconstruction of Neurons with Flood-filling Networks
Authors Michał Januszewski, Jörgen Kornfeld, Peter H. Li, Art Pope, Tim Blakely, Larry Lindsey, Jeremy Maitin-Shepard, Mike Tyka, Winfried Denk, Viren Jain
Abstract Reconstruction of neural circuits from volume electron microscopy data requires the tracing of complete cells including all their neurites. Automated approaches have been developed to perform the tracing, but without costly human proofreading their error rates are too high to obtain reliable circuit diagrams. We present a method for automated segmentation that, like the majority of previous efforts, employs convolutional neural networks, but contains in addition a recurrent pathway that allows the iterative optimization and extension of the reconstructed shape of individual neural processes. We used this technique, which we call flood-filling networks, to trace neurons in a data set obtained by serial block-face electron microscopy from a male zebra finch brain. Our method achieved a mean error-free neurite path length of 1.1 mm, an order of magnitude better than previously published approaches applied to the same dataset. Only 4 mergers were observed in a neurite test set of 97 mm path length.
Tasks Electron Microscopy Image Segmentation, Image Reconstruction
Published 2017-10-09
URL https://doi.org/10.1038/s41592-018-0049-4
PDF https://www.biorxiv.org/content/biorxiv/early/2017/10/09/200675.full-text.pdf
PWC https://paperswithcode.com/paper/high-precision-automated-reconstruction-of
Repo
Framework

以知識表徵方法建構台語聲調群剖析器 (A Knowledge Representation Method to Implement A Taiwanese Tone Group Parser) [In Chinese]

Title 以知識表徵方法建構台語聲調群剖析器 (A Knowledge Representation Method to Implement A Taiwanese Tone Group Parser) [In Chinese]
Authors Yu-Chu Chang
Abstract
Tasks
Published 2017-11-01
URL https://www.aclweb.org/anthology/O17-1001/
PDF https://www.aclweb.org/anthology/O17-1001
PWC https://paperswithcode.com/paper/acee-a3413aoaeae2eac34aa-a-knowledge-1
Repo
Framework

Recycling Privileged Learning and Distribution Matching for Fairness

Title Recycling Privileged Learning and Distribution Matching for Fairness
Authors Novi Quadrianto, Viktoriia Sharmanska
Abstract Equipping machine learning models with ethical and legal constraints is a serious issue; without this, the future of machine learning is at risk. This paper takes a step forward in this direction and focuses on ensuring machine learning models deliver fair decisions. In legal scholarships, the notion of fairness itself is evolving and multi-faceted. We set an overarching goal to develop a unified machine learning framework that is able to handle any definitions of fairness, their combinations, and also new definitions that might be stipulated in the future. To achieve our goal, we recycle two well-established machine learning techniques, privileged learning and distribution matching, and harmonize them for satisfying multi-faceted fairness definitions. We consider protected characteristics such as race and gender as privileged information that is available at training but not at test time; this accelerates model training and delivers fairness through unawareness. Further, we cast demographic parity, equalized odds, and equality of opportunity as a classical two-sample problem of conditional distributions, which can be solved in a general form by using distance measures in Hilbert Space. We show several existing models are special cases of ours. Finally, we advocate returning the Pareto frontier of multi-objective minimization of error and unfairness in predictions. This will facilitate decision makers to select an operating point and to be accountable for it.
Tasks
Published 2017-12-01
URL http://papers.nips.cc/paper/6670-recycling-privileged-learning-and-distribution-matching-for-fairness
PDF http://papers.nips.cc/paper/6670-recycling-privileged-learning-and-distribution-matching-for-fairness.pdf
PWC https://paperswithcode.com/paper/recycling-privileged-learning-and
Repo
Framework

BUSEM at SemEval-2017 Task 4A Sentiment Analysis with Word Embedding and Long Short Term Memory RNN Approaches

Title BUSEM at SemEval-2017 Task 4A Sentiment Analysis with Word Embedding and Long Short Term Memory RNN Approaches
Authors Deger Ayata, Murat Saraclar, Arzucan Ozgur
Abstract This paper describes our approach for SemEval-2017 Task 4: Sentiment Analysis in Twitter. We have participated in Subtask A: Message Polarity Classification subtask and developed two systems. The first system uses word embeddings for feature representation and Support Vector Machine, Random Forest and Naive Bayes algorithms for classification of Twitter messages into negative, neutral and positive polarity. The second system is based on Long Short Term Memory Recurrent Neural Networks and uses word indexes as sequence of inputs for feature representation.
Tasks Opinion Mining, Sentiment Analysis, Subjectivity Analysis, Word Embeddings
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2131/
PDF https://www.aclweb.org/anthology/S17-2131
PWC https://paperswithcode.com/paper/busem-at-semeval-2017-task-4a-sentiment
Repo
Framework

Enhancing Drug-Drug Interaction Classification with Corpus-level Feature and Classifier Ensemble

Title Enhancing Drug-Drug Interaction Classification with Corpus-level Feature and Classifier Ensemble
Authors Jing Cyun Tu, Po-Ting Lai, Richard Tzong-Han Tsai
Abstract The study of drug-drug interaction (DDI) is important in the drug discovering. Both PubMed and DrugBank are rich resources to retrieve DDI information which is usually represented in plain text. Automatically extracting DDI pairs from text improves the quality of drug discov-ering. In this paper, we presented a study that focuses on the DDI classification. We normalized the drug names, and developed both sentence-level and corpus-level features for DDI classification. A classifier ensemble approach is used for the unbalance DDI labels problem. Our approach achieved an F-score of 65.4{%} on SemEval 2013 DDI test set. The experimental results also show the effects of proposed corpus-level features in the DDI task.
Tasks Feature Engineering, Named Entity Recognition
Published 2017-11-01
URL https://www.aclweb.org/anthology/W17-5808/
PDF https://www.aclweb.org/anthology/W17-5808
PWC https://paperswithcode.com/paper/enhancing-drug-drug-interaction
Repo
Framework

Classifying Frames at the Sentence Level in News Articles

Title Classifying Frames at the Sentence Level in News Articles
Authors Nona Naderi, Graeme Hirst
Abstract Previous approaches to generic frame classification analyze frames at the document level. Here, we propose a supervised based approach based on deep neural networks and distributional representations for classifying frames at the sentence level in news articles. We conduct our experiments on the publicly available Media Frames Corpus compiled from the U.S. Newspapers. Using (B)LSTMs and GRU networks to represent the meaning of frames, we demonstrate that our approach yields at least 14-point improvement over several baseline methods.
Tasks Semantic Textual Similarity, Time Series, Topic Models
Published 2017-09-01
URL https://www.aclweb.org/anthology/R17-1070/
PDF https://doi.org/10.26615/978-954-452-049-6_070
PWC https://paperswithcode.com/paper/classifying-frames-at-the-sentence-level-in
Repo
Framework
comments powered by Disqus