July 26, 2019

1853 words 9 mins read

Paper Group NANR 88

Paper Group NANR 88

Proceedings of the Fifth International Workshop on Natural Language Processing for Social Media. Verb Replacer: An English Verb Error Correction System. Investigating Diatopic Variation in a Historical Corpus. A simple model of recognition and recall memory. Inference is Everything: Recasting Semantic Resources into a Unified Evaluation Framework. …

Proceedings of the Fifth International Workshop on Natural Language Processing for Social Media

Title Proceedings of the Fifth International Workshop on Natural Language Processing for Social Media
Authors
Abstract
Tasks
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-1100/
PDF https://www.aclweb.org/anthology/W17-1100
PWC https://paperswithcode.com/paper/proceedings-of-the-fifth-international
Repo
Framework

Verb Replacer: An English Verb Error Correction System

Title Verb Replacer: An English Verb Error Correction System
Authors Yu-Hsuan Wu, Jhih-Jie Chen, Jason Chang
Abstract According to the analysis of Cambridge Learner Corpus, using a wrong verb is the most common type of grammatical errors. This paper describes Verb Replacer, a system for detecting and correcting potential verb errors in a given sentence. In our approach, alternative verbs are considered to replace the verb based on an error-annotated corpus and verb-object collocations. The method involves applying regression on channel models, parsing the sentence, identifying the verbs, retrieving a small set of alternative verbs, and evaluating each alternative. Our method combines and improves channel and language models, resulting in high recall of detecting and correcting verb misuse.
Tasks
Published 2017-11-01
URL https://www.aclweb.org/anthology/I17-3013/
PDF https://www.aclweb.org/anthology/I17-3013
PWC https://paperswithcode.com/paper/verb-replacer-an-english-verb-error
Repo
Framework

Investigating Diatopic Variation in a Historical Corpus

Title Investigating Diatopic Variation in a Historical Corpus
Authors Stefanie Dipper, S Waldenberger, ra
Abstract This paper investigates diatopic variation in a historical corpus of German. Based on equivalent word forms from different language areas, replacement rules and mappings are derived which describe the relations between these word forms. These rules and mappings are then interpreted as reflections of morphological, phonological or graphemic variation. Based on sample rules and mappings, we show that our approach can replicate results from historical linguistics. While previous studies were restricted to predefined word lists, or confined to single authors or texts, our approach uses a much wider range of data available in historical corpora.
Tasks
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-1204/
PDF https://www.aclweb.org/anthology/W17-1204
PWC https://paperswithcode.com/paper/investigating-diatopic-variation-in-a
Repo
Framework

A simple model of recognition and recall memory

Title A simple model of recognition and recall memory
Authors Nisheeth Srivastava, Edward Vul
Abstract We show that several striking differences in memory performance between recognition and recall tasks are explained by an ecological bias endemic in classic memory experiments - that such experiments universally involve more stimuli than retrieval cues. We show that while it is sensible to think of recall as simply retrieving items when probed with a cue - typically the item list itself - it is better to think of recognition as retrieving cues when probed with items. To test this theory, by manipulating the number of items and cues in a memory experiment, we show a crossover effect in memory performance within subjects such that recognition performance is superior to recall performance when the number of items is greater than the number of cues and recall performance is better than recognition when the converse holds. We build a simple computational model around this theory, using sampling to approximate an ideal Bayesian observer encoding and retrieving situational co-occurrence frequencies of stimuli and retrieval cues. This model robustly reproduces a number of dissociations in recognition and recall previously used to argue for dual-process accounts of declarative memory.
Tasks
Published 2017-12-01
URL http://papers.nips.cc/paper/6633-a-simple-model-of-recognition-and-recall-memory
PDF http://papers.nips.cc/paper/6633-a-simple-model-of-recognition-and-recall-memory.pdf
PWC https://paperswithcode.com/paper/a-simple-model-of-recognition-and-recall
Repo
Framework

Inference is Everything: Recasting Semantic Resources into a Unified Evaluation Framework

Title Inference is Everything: Recasting Semantic Resources into a Unified Evaluation Framework
Authors Aaron Steven White, Pushpendre Rastogi, Kevin Duh, Benjamin Van Durme
Abstract We propose to unify a variety of existing semantic classification tasks, such as semantic role labeling, anaphora resolution, and paraphrase detection, under the heading of Recognizing Textual Entailment (RTE). We present a general strategy to automatically generate one or more sentential hypotheses based on an input sentence and pre-existing manual semantic annotations. The resulting suite of datasets enables us to probe a statistical RTE model{'}s performance on different aspects of semantics. We demonstrate the value of this approach by investigating the behavior of a popular neural network RTE model.
Tasks Image Captioning, Natural Language Inference, Semantic Role Labeling
Published 2017-11-01
URL https://www.aclweb.org/anthology/I17-1100/
PDF https://www.aclweb.org/anthology/I17-1100
PWC https://paperswithcode.com/paper/inference-is-everything-recasting-semantic
Repo
Framework

Towards Debate Automation: a Recurrent Model for Predicting Debate Winners

Title Towards Debate Automation: a Recurrent Model for Predicting Debate Winners
Authors Peter Potash, Anna Rumshisky
Abstract In this paper we introduce a practical first step towards the creation of an automated debate agent: a state-of-the-art recurrent predictive model for predicting debate winners. By having an accurate predictive model, we are able to objectively rate the quality of a statement made at a specific turn in a debate. The model is based on a recurrent neural network architecture with attention, which allows the model to effectively account for the entire debate when making its prediction. Our model achieves state-of-the-art accuracy on a dataset of debate transcripts annotated with audience favorability of the debate teams. Finally, we discuss how future work can leverage our proposed model for the creation of an automated debate agent. We accomplish this by determining the model input that will maximize audience favorability toward a given side of a debate at an arbitrary turn.
Tasks Text Generation
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1261/
PDF https://www.aclweb.org/anthology/D17-1261
PWC https://paperswithcode.com/paper/towards-debate-automation-a-recurrent-model
Repo
Framework

ADAPT Centre Cone Team at IJCNLP-2017 Task 5: A Similarity-Based Logistic Regression Approach to Multi-choice Question Answering in an Examinations Shared Task

Title ADAPT Centre Cone Team at IJCNLP-2017 Task 5: A Similarity-Based Logistic Regression Approach to Multi-choice Question Answering in an Examinations Shared Task
Authors Daria Dzendzik, Alberto Poncelas, Carl Vogel, Qun Liu
Abstract We describe the work of a team from the ADAPT Centre in Ireland in addressing automatic answer selection for the Multi-choice Question Answering in Examinations shared task. The system is based on a logistic regression over the string similarities between question, answer, and additional text. We obtain the highest grade out of six systems: 48.7{%} accuracy on a validation set (vs. a baseline of 29.45{%}) and 45.6{%} on a test set.
Tasks Answer Selection, Question Answering
Published 2017-12-01
URL https://www.aclweb.org/anthology/I17-4010/
PDF https://www.aclweb.org/anthology/I17-4010
PWC https://paperswithcode.com/paper/adapt-centre-cone-team-at-ijcnlp-2017-task-5
Repo
Framework

Conjunctive Categorial Grammars

Title Conjunctive Categorial Grammars
Authors Stepan Kuznetsov, Alex Okhotin, er
Abstract
Tasks
Published 2017-07-01
URL https://www.aclweb.org/anthology/W17-3414/
PDF https://www.aclweb.org/anthology/W17-3414
PWC https://paperswithcode.com/paper/conjunctive-categorial-grammars
Repo
Framework

DataStories at SemEval-2017 Task 6: Siamese LSTM with Attention for Humorous Text Comparison

Title DataStories at SemEval-2017 Task 6: Siamese LSTM with Attention for Humorous Text Comparison
Authors Christos Baziotis, Nikos Pelekis, Christos Doulkeridis
Abstract In this paper we present a deep-learning system that competed at SemEval-2017 Task 6 ‘'{#}HashtagWars: Learning a Sense of Humor{''}. We participated in Subtask A, in which the goal was, given two Twitter messages, to identify which one is funnier. We propose a Siamese architecture with bidirectional Long Short-Term Memory (LSTM) networks, augmented with an attention mechanism. Our system works on the token-level, leveraging word embeddings trained on a big collection of unlabeled Twitter messages. We ranked 2nd in 7 teams. A post-completion improvement of our model, achieves state-of-the-art results on {#}HashtagWars dataset.
Tasks Feature Engineering, Humor Detection, Word Embeddings
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2065/
PDF https://www.aclweb.org/anthology/S17-2065
PWC https://paperswithcode.com/paper/datastories-at-semeval-2017-task-6-siamese
Repo
Framework

Adullam at SemEval-2017 Task 4: Sentiment Analyzer Using Lexicon Integrated Convolutional Neural Networks with Attention

Title Adullam at SemEval-2017 Task 4: Sentiment Analyzer Using Lexicon Integrated Convolutional Neural Networks with Attention
Authors Joosung Yoon, Kigon Lyu, Hyeoncheol Kim
Abstract We propose a sentiment analyzer for the prediction of document-level sentiments of English micro-blog messages from Twitter. The proposed method is based on lexicon integrated convolutional neural networks with attention (LCA). Its performance was evaluated using the datasets provided by SemEval competition (Task 4). The proposed sentiment analyzer obtained an average F1 of 55.2{%}, an average recall of 58.9{%} and an accuracy of 61.4{%}.
Tasks Sentiment Analysis, Word Embeddings
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2123/
PDF https://www.aclweb.org/anthology/S17-2123
PWC https://paperswithcode.com/paper/adullam-at-semeval-2017-task-4-sentiment
Repo
Framework

Word-Context Character Embeddings for Chinese Word Segmentation

Title Word-Context Character Embeddings for Chinese Word Segmentation
Authors Hao Zhou, Zhenting Yu, Yue Zhang, Shujian Huang, Xinyu Dai, Jiajun Chen
Abstract Neural parsers have benefited from automatically labeled data via dependency-context word embeddings. We investigate training character embeddings on a word-based context in a similar way, showing that the simple method improves state-of-the-art neural word segmentation models significantly, beating tri-training baselines for leveraging auto-segmented data.
Tasks Chinese Word Segmentation, Domain Adaptation, Word Embeddings
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1079/
PDF https://www.aclweb.org/anthology/D17-1079
PWC https://paperswithcode.com/paper/word-context-character-embeddings-for-chinese
Repo
Framework

Lexical Disambiguation of Igbo using Diacritic Restoration

Title Lexical Disambiguation of Igbo using Diacritic Restoration
Authors Ignatius Ezeani, Mark Hepple, Ikechukwu Onyenwe
Abstract Properly written texts in Igbo, a low-resource African language, are rich in both orthographic and tonal diacritics. Diacritics are essential in capturing the distinctions in pronunciation and meaning of words, as well as in lexical disambiguation. Unfortunately, most electronic texts in diacritic languages are written without diacritics. This makes diacritic restoration a necessary step in corpus building and language processing tasks for languages with diacritics. In our previous work, we built some n-gram models with simple smoothing techniques based on a closed-world assumption. However, as a classification task, diacritic restoration is well suited for and will be more generalisable with machine learning. This paper, therefore, presents a more standard approach to dealing with the task which involves the application of machine learning algorithms.
Tasks
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-1907/
PDF https://www.aclweb.org/anthology/W17-1907
PWC https://paperswithcode.com/paper/lexical-disambiguation-of-igbo-using
Repo
Framework

RTM at SemEval-2017 Task 1: Referential Translation Machines for Predicting Semantic Similarity

Title RTM at SemEval-2017 Task 1: Referential Translation Machines for Predicting Semantic Similarity
Authors Ergun Bi{\c{c}}ici
Abstract We use referential translation machines for predicting the semantic similarity of text in all STS tasks which contain Arabic, English, Spanish, and Turkish this year. RTMs pioneer a language independent approach to semantic similarity and remove the need to access any task or domain specific information or resource. RTMs become 6th out of 52 submissions in Spanish to English STS. We average prediction scores using weights based on the training performance to improve the overall performance.
Tasks Feature Selection, Machine Translation, Semantic Similarity, Semantic Textual Similarity
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2030/
PDF https://www.aclweb.org/anthology/S17-2030
PWC https://paperswithcode.com/paper/rtm-at-semeval-2017-task-1-referential
Repo
Framework

Ensemble Methods for Native Language Identification

Title Ensemble Methods for Native Language Identification
Authors Sophia Chan, Maryam Honari Jahromi, Benjamin Benetti, Aazim Lakhani, Alona Fyshe
Abstract Our team{—}Uvic-NLP{—}explored and evaluated a variety of lexical features for Native Language Identification (NLI) within the framework of ensemble methods. Using a subset of the highest performing features, we train Support Vector Machines (SVM) and Fully Connected Neural Networks (FCNN) as base classifiers, and test different methods for combining their outputs. Restricting our scope to the closed essay track in the NLI Shared Task 2017, we find that our best SVM ensemble achieves an F1 score of 0.8730 on the test set.
Tasks Language Acquisition, Language Identification, Native Language Identification
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-5023/
PDF https://www.aclweb.org/anthology/W17-5023
PWC https://paperswithcode.com/paper/ensemble-methods-for-native-language
Repo
Framework

Coarticulatory propensity in Khalkha Mongolian

Title Coarticulatory propensity in Khalkha Mongolian
Authors Ushashi Banerjee, Indranil Dutta, Irfan S
Abstract
Tasks
Published 2017-12-01
URL https://www.aclweb.org/anthology/W17-7544/
PDF https://www.aclweb.org/anthology/W17-7544
PWC https://paperswithcode.com/paper/coarticulatory-propensity-in-khalkha
Repo
Framework
comments powered by Disqus