July 26, 2019

1702 words 8 mins read

Paper Group NANR 47

Paper Group NANR 47

Supersense Tagging with a Combination of Character, Subword, and Word-level Representations. HistoBankVis: Detecting Language Change via Data Visualization. Improving Word Sense Disambiguation in Neural Machine Translation with Sense Embeddings. UDLex: Towards Cross-language Subcategorization Lexicons. Learning an Input Filter for Argument Structur …

Supersense Tagging with a Combination of Character, Subword, and Word-level Representations

Title Supersense Tagging with a Combination of Character, Subword, and Word-level Representations
Authors Youhyun Shin, Sang-goo Lee
Abstract Recently, there has been increased interest in utilizing characters or subwords for natural language processing (NLP) tasks. However, the effect of utilizing character, subword, and word-level information simultaneously has not been examined so far. In this paper, we propose a model to leverage various levels of input features to improve on the performance of an supersense tagging task. Detailed analysis of experimental results show that different levels of input representation offer distinct characteristics that explain performance discrepancy among different tasks.
Tasks Entity Alignment, Language Modelling, Machine Translation, Named Entity Recognition, Word Sense Disambiguation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4106/
PDF https://www.aclweb.org/anthology/W17-4106
PWC https://paperswithcode.com/paper/supersense-tagging-with-a-combination-of
Repo
Framework

HistoBankVis: Detecting Language Change via Data Visualization

Title HistoBankVis: Detecting Language Change via Data Visualization
Authors Christin Sch{"a}tzle, Michael Hund, Frederik Dennig, Miriam Butt, Daniel Keim
Abstract
Tasks
Published 2017-05-01
URL https://www.aclweb.org/anthology/W17-0507/
PDF https://www.aclweb.org/anthology/W17-0507
PWC https://paperswithcode.com/paper/histobankvis-detecting-language-change-via
Repo
Framework

Improving Word Sense Disambiguation in Neural Machine Translation with Sense Embeddings

Title Improving Word Sense Disambiguation in Neural Machine Translation with Sense Embeddings
Authors Annette Rios Gonzales, Laura Mascarell, Rico Sennrich
Abstract
Tasks Machine Translation, Word Sense Disambiguation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4702/
PDF https://www.aclweb.org/anthology/W17-4702
PWC https://paperswithcode.com/paper/improving-word-sense-disambiguation-in-neural
Repo
Framework

UDLex: Towards Cross-language Subcategorization Lexicons

Title UDLex: Towards Cross-language Subcategorization Lexicons
Authors Giulia Rambelli, Aless Lenci, ro, Thierry Poibeau
Abstract
Tasks Information Retrieval, Machine Translation, Natural Language Inference, Word Sense Disambiguation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-6524/
PDF https://www.aclweb.org/anthology/W17-6524
PWC https://paperswithcode.com/paper/udlex-towards-cross-language
Repo
Framework

Learning an Input Filter for Argument Structure Acquisition

Title Learning an Input Filter for Argument Structure Acquisition
Authors Laurel Perkins, Naomi Feldman, Jeffrey Lidz
Abstract How do children learn a verb{'}s argument structure when their input contains nonbasic clauses that obscure verb transitivity? Here we present a new model that infers verb transitivity by learning to filter out non-basic clauses that were likely parsed in error. In simulations with child-directed speech, we show that this model accurately categorizes the majority of 50 frequent transitive, intransitive and alternating verbs, and jointly learns appropriate parameters for filtering parsing errors. Our model is thus able to filter out problematic data for verb learning without knowing in advance which data need to be filtered.
Tasks
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-0702/
PDF https://www.aclweb.org/anthology/W17-0702
PWC https://paperswithcode.com/paper/learning-an-input-filter-for-argument
Repo
Framework

Oxford at SemEval-2017 Task 9: Neural AMR Parsing with Pointer-Augmented Attention

Title Oxford at SemEval-2017 Task 9: Neural AMR Parsing with Pointer-Augmented Attention
Authors Jan Buys, Phil Blunsom
Abstract We present a neural encoder-decoder AMR parser that extends an attention-based model by predicting the alignment between graph nodes and sentence tokens explicitly with a pointer mechanism. Candidate lemmas are predicted as a pre-processing step so that the lemmas of lexical concepts, as well as constant strings, are factored out of the graph linearization and recovered through the predicted alignments. The approach does not rely on syntactic parses or extensive external resources. Our parser obtained 59{%} Smatch on the SemEval test set.
Tasks Amr Parsing, Lemmatization
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2157/
PDF https://www.aclweb.org/anthology/S17-2157
PWC https://paperswithcode.com/paper/oxford-at-semeval-2017-task-9-neural-amr
Repo
Framework

A Sharp Error Analysis for the Fused Lasso, with Application to Approximate Changepoint Screening

Title A Sharp Error Analysis for the Fused Lasso, with Application to Approximate Changepoint Screening
Authors Kevin Lin, James L. Sharpnack, Alessandro Rinaldo, Ryan J. Tibshirani
Abstract In the 1-dimensional multiple changepoint detection problem, we derive a new fast error rate for the fused lasso estimator, under the assumption that the mean vector has a sparse number of changepoints. This rate is seen to be suboptimal (compared to the minimax rate) by only a factor of $\log\log{n}$. Our proof technique is centered around a novel construction that we call a lower interpolant. We extend our results to misspecified models and exponential family distributions. We also describe the implications of our error analysis for the approximate screening of changepoints.
Tasks
Published 2017-12-01
URL http://papers.nips.cc/paper/7264-a-sharp-error-analysis-for-the-fused-lasso-with-application-to-approximate-changepoint-screening
PDF http://papers.nips.cc/paper/7264-a-sharp-error-analysis-for-the-fused-lasso-with-application-to-approximate-changepoint-screening.pdf
PWC https://paperswithcode.com/paper/a-sharp-error-analysis-for-the-fused-lasso
Repo
Framework

A Short Survey on Taxonomy Learning from Text Corpora: Issues, Resources and Recent Advances

Title A Short Survey on Taxonomy Learning from Text Corpora: Issues, Resources and Recent Advances
Authors Chengyu Wang, Xiaofeng He, Aoying Zhou
Abstract A taxonomy is a semantic hierarchy, consisting of concepts linked by is-a relations. While a large number of taxonomies have been constructed from human-compiled resources (e.g., Wikipedia), learning taxonomies from text corpora has received a growing interest and is essential for long-tailed and domain-specific knowledge acquisition. In this paper, we overview recent advances on taxonomy construction from free texts, reorganizing relevant subtasks into a complete framework. We also overview resources for evaluation and discuss challenges for future research.
Tasks Question Answering
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1123/
PDF https://www.aclweb.org/anthology/D17-1123
PWC https://paperswithcode.com/paper/a-short-survey-on-taxonomy-learning-from-text
Repo
Framework

Annotating Speech, Attitude and Perception Reports

Title Annotating Speech, Attitude and Perception Reports
Authors Corien Bary, Leopold Hess, Kees Thijs, Peter Berck, Iris Hendrickx
Abstract We present REPORTS, an annotation scheme for the annotation of speech, attitude and perception reports. Such a scheme makes it possible to annotate the various text elements involved in such reports (e.g. embedding entity, complement, complement head) and their relations in a uniform way, which in turn facilitates the automatic extraction of information on, for example, complementation and vocabulary distribution. We also present the Ancient Greek corpus RAG (Thucydides{'} History of the Peloponnesian War), to which we have applied this scheme using the annotation tool BRAT. We discuss some of the issues, both theoretical and practical, that we encountered, show how the corpus helps in answering specific questions, and conclude that REPORTS fitted in well with our needs.
Tasks
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-0806/
PDF https://www.aclweb.org/anthology/W17-0806
PWC https://paperswithcode.com/paper/annotating-speech-attitude-and-perception
Repo
Framework

Swedish Prepositions are not Pure Function Words

Title Swedish Prepositions are not Pure Function Words
Authors Lars Ahrenberg
Abstract
Tasks
Published 2017-05-01
URL https://www.aclweb.org/anthology/W17-0402/
PDF https://www.aclweb.org/anthology/W17-0402
PWC https://paperswithcode.com/paper/swedish-prepositions-are-not-pure-function
Repo
Framework

Can Syntax Help? Improving an LSTM-based Sentence Compression Model for New Domains

Title Can Syntax Help? Improving an LSTM-based Sentence Compression Model for New Domains
Authors Liangguo Wang, Jing Jiang, Hai Leong Chieu, Chen Hui Ong, D Song, an, Lejian Liao
Abstract In this paper, we study how to improve the domain adaptability of a deletion-based Long Short-Term Memory (LSTM) neural network model for sentence compression. We hypothesize that syntactic information helps in making such models more robust across domains. We propose two major changes to the model: using explicit syntactic features and introducing syntactic constraints through Integer Linear Programming (ILP). Our evaluation shows that the proposed model works better than the original model as well as a traditional non-neural-network-based model in a cross-domain setting.
Tasks Sentence Compression, Text Summarization, Tokenization
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-1127/
PDF https://www.aclweb.org/anthology/P17-1127
PWC https://paperswithcode.com/paper/can-syntax-help-improving-an-lstm-based
Repo
Framework

Word Similarity Datasets for Indian Languages: Annotation and Baseline Systems

Title Word Similarity Datasets for Indian Languages: Annotation and Baseline Systems
Authors Syed Sarfaraz Akhtar, Arihant Gupta, Avijit Vajpayee, Arjit Srivastava, Manish Shrivastava
Abstract With the advent of word representations, word similarity tasks are becoming increasing popular as an evaluation metric for the quality of the representations. In this paper, we present manually annotated monolingual word similarity datasets of six Indian languages - Urdu, Telugu, Marathi, Punjabi, Tamil and Gujarati. These languages are most spoken Indian languages worldwide after Hindi and Bengali. For the construction of these datasets, our approach relies on translation and re-annotation of word similarity datasets of English. We also present baseline scores for word representation models using state-of-the-art techniques for Urdu, Telugu and Marathi by evaluating them on newly created word similarity datasets.
Tasks Dependency Parsing, Machine Translation, Named Entity Recognition, Question Answering, Semantic Textual Similarity, Word Sense Disambiguation
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-0811/
PDF https://www.aclweb.org/anthology/W17-0811
PWC https://paperswithcode.com/paper/word-similarity-datasets-for-indian-languages
Repo
Framework

Spatial-Semantic Image Search by Visual Feature Synthesis

Title Spatial-Semantic Image Search by Visual Feature Synthesis
Authors Long Mai, Hailin Jin, Zhe Lin, Chen Fang, Jonathan Brandt, Feng Liu
Abstract The performance of image retrieval has been improved tremendously in recent years through the use of deep feature representations. Most existing methods, however, aim to retrieve images that are visually similar or semantically relevant to the query, irrespective of spatial configuration. In this paper, we develop a spatial-semantic image search technology that enables users to search for images with both semantic and spatial constraints by manipulating concept text-boxes on a 2D query canvas. We train a convolutional neural network to synthesize appropriate visual features that captures the spatial-semantic constraints from the user canvas query. We directly optimize the retrieval performance of the visual features when training our deep neural network. These visual features then are used to retrieve images that are both spatially and semantically relevant to the user query. The experiments on large-scale datasets such as MS-COCO and Visual Genome show that our method outperforms other baseline and state-of-the-art methods in spatial-semantic image search.
Tasks Image Retrieval
Published 2017-07-01
URL http://openaccess.thecvf.com/content_cvpr_2017/html/Mai_Spatial-Semantic_Image_Search_CVPR_2017_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2017/papers/Mai_Spatial-Semantic_Image_Search_CVPR_2017_paper.pdf
PWC https://paperswithcode.com/paper/spatial-semantic-image-search-by-visual
Repo
Framework

When Conditional Logic met Connexive Logic

Title When Conditional Logic met Connexive Logic
Authors Mathieu Vidal
Abstract
Tasks
Published 2017-01-01
URL https://www.aclweb.org/anthology/W17-6816/
PDF https://www.aclweb.org/anthology/W17-6816
PWC https://paperswithcode.com/paper/when-conditional-logic-met-connexive-logic
Repo
Framework

Earth Mover’s Distance Minimization for Unsupervised Bilingual Lexicon Induction

Title Earth Mover’s Distance Minimization for Unsupervised Bilingual Lexicon Induction
Authors Meng Zhang, Yang Liu, Huanbo Luan, Maosong Sun
Abstract Cross-lingual natural language processing hinges on the premise that there exists invariance across languages. At the word level, researchers have identified such invariance in the word embedding semantic spaces of different languages. However, in order to connect the separate spaces, cross-lingual supervision encoded in parallel data is typically required. In this paper, we attempt to establish the cross-lingual connection without relying on any cross-lingual supervision. By viewing word embedding spaces as distributions, we propose to minimize their earth mover{'}s distance, a measure of divergence between distributions. We demonstrate the success on the unsupervised bilingual lexicon induction task. In addition, we reveal an interesting finding that the earth mover{'}s distance shows potential as a measure of language difference.
Tasks Cross-Lingual Transfer, Word Embeddings
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1207/
PDF https://www.aclweb.org/anthology/D17-1207
PWC https://paperswithcode.com/paper/earth-movers-distance-minimization-for
Repo
Framework
comments powered by Disqus