July 26, 2019

2028 words 10 mins read

Paper Group NANR 96

Paper Group NANR 96

Maximizing Subset Accuracy with Recurrent Neural Networks in Multi-label Classification. Accent Adaptation for the Air Traffic Control Domain. Neural Graphical Models over Strings for Principal Parts Morphological Paradigm Completion. Pulling Out the Stops: Rethinking Stopword Removal for Topic Models. Multimodal Topic Labelling. Bib2vec: Embedding …

Maximizing Subset Accuracy with Recurrent Neural Networks in Multi-label Classification

Title Maximizing Subset Accuracy with Recurrent Neural Networks in Multi-label Classification
Authors Jinseok Nam, Eneldo Loza Mencía, Hyunwoo J. Kim, Johannes Fürnkranz
Abstract Multi-label classification is the task of predicting a set of labels for a given input instance. Classifier chains are a state-of-the-art method for tackling such problems, which essentially converts this problem into a sequential prediction problem, where the labels are first ordered in an arbitrary fashion, and the task is to predict a sequence of binary values for these labels. In this paper, we replace classifier chains with recurrent neural networks, a sequence-to-sequence prediction algorithm which has recently been successfully applied to sequential prediction tasks in many domains. The key advantage of this approach is that it allows to focus on the prediction of the positive labels only, a much smaller set than the full set of possible labels. Moreover, parameter sharing across all classifiers allows to better exploit information of previous decisions. As both, classifier chains and recurrent neural networks depend on a fixed ordering of the labels, which is typically not part of a multi-label problem specification, we also compare different ways of ordering the label set, and give some recommendations on suitable ordering strategies.
Tasks Multi-Label Classification
Published 2017-12-01
URL http://papers.nips.cc/paper/7125-maximizing-subset-accuracy-with-recurrent-neural-networks-in-multi-label-classification
PDF http://papers.nips.cc/paper/7125-maximizing-subset-accuracy-with-recurrent-neural-networks-in-multi-label-classification.pdf
PWC https://paperswithcode.com/paper/maximizing-subset-accuracy-with-recurrent
Repo
Framework

Accent Adaptation for the Air Traffic Control Domain

Title Accent Adaptation for the Air Traffic Control Domain
Authors Matthew Garber, Meital Singer, Christopher Ward
Abstract
Tasks Speech Recognition
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-3016/
PDF https://www.aclweb.org/anthology/P17-3016
PWC https://paperswithcode.com/paper/accent-adaptation-for-the-air-traffic-control
Repo
Framework

Neural Graphical Models over Strings for Principal Parts Morphological Paradigm Completion

Title Neural Graphical Models over Strings for Principal Parts Morphological Paradigm Completion
Authors Ryan Cotterell, John Sylak-Glassman, Christo Kirov
Abstract Many of the world{'}s languages contain an abundance of inflected forms for each lexeme. A critical task in processing such languages is predicting these inflected forms. We develop a novel statistical model for the problem, drawing on graphical modeling techniques and recent advances in deep learning. We derive a Metropolis-Hastings algorithm to jointly decode the model. Our Bayesian network draws inspiration from principal parts morphological analysis. We demonstrate improvements on 5 languages.
Tasks Morphological Analysis
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-2120/
PDF https://www.aclweb.org/anthology/E17-2120
PWC https://paperswithcode.com/paper/neural-graphical-models-over-strings-for
Repo
Framework

Pulling Out the Stops: Rethinking Stopword Removal for Topic Models

Title Pulling Out the Stops: Rethinking Stopword Removal for Topic Models
Authors Alex Schofield, ra, M{\aa}ns Magnusson, David Mimno
Abstract It is often assumed that topic models benefit from the use of a manually curated stopword list. Constructing this list is time-consuming and often subject to user judgments about what kinds of words are important to the model and the application. Although stopword removal clearly affects which word types appear as most probable terms in topics, we argue that this improvement is superficial, and that topic inference benefits little from the practice of removing stopwords beyond very frequent terms. Removing corpus-specific stopwords after model inference is more transparent and produces similar results to removing those words prior to inference.
Tasks Language Modelling, Topic Models
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-2069/
PDF https://www.aclweb.org/anthology/E17-2069
PWC https://paperswithcode.com/paper/pulling-out-the-stops-rethinking-stopword
Repo
Framework

Multimodal Topic Labelling

Title Multimodal Topic Labelling
Authors Ionut Sorodoc, Jey Han Lau, Nikolaos Aletras, Timothy Baldwin
Abstract Topics generated by topic models are typically presented as a list of topic terms. Automatic topic labelling is the task of generating a succinct label that summarises the theme or subject of a topic, with the intention of reducing the cognitive load of end-users when interpreting these topics. Traditionally, topic label systems focus on a single label modality, e.g. textual labels. In this work we propose a multimodal approach to topic labelling using a simple feedforward neural network. Given a topic and a candidate image or textual label, our method automatically generates a rating for the label, relative to the topic. Experiments show that this multimodal approach outperforms single-modality topic labelling systems.
Tasks Topic Models
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-2111/
PDF https://www.aclweb.org/anthology/E17-2111
PWC https://paperswithcode.com/paper/multimodal-topic-labelling
Repo
Framework

Bib2vec: Embedding-based Search System for Bibliographic Information

Title Bib2vec: Embedding-based Search System for Bibliographic Information
Authors Takuma Yoneda, Koki Mori, Makoto Miwa, Yutaka Sasaki
Abstract We propose a novel embedding model that represents relationships among several elements in bibliographic information with high representation ability and flexibility. Based on this model, we present a novel search system that shows the relationships among the elements in the ACL Anthology Reference Corpus. The evaluation results show that our model can achieve a high prediction ability and produce reasonable search results.
Tasks Network Embedding, Topic Models
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-3028/
PDF https://www.aclweb.org/anthology/E17-3028
PWC https://paperswithcode.com/paper/bib2vec-embedding-based-search-system-for
Repo
Framework

Improving Verb Metaphor Detection by Propagating Abstractness to Words, Phrases and Individual Senses

Title Improving Verb Metaphor Detection by Propagating Abstractness to Words, Phrases and Individual Senses
Authors Maximilian K{"o}per, Sabine Schulte im Walde
Abstract Abstract words refer to things that can not be seen, heard, felt, smelled, or tasted as opposed to concrete words. Among other applications, the degree of abstractness has been shown to be a useful information for metaphor detection. Our contribution to this topic are as follows: i) we compare supervised techniques to learn and extend abstractness ratings for huge vocabularies ii) we learn and investigate norms for larger units by propagating abstractness to verb-noun pairs which lead to better metaphor detection iii) we overcome the limitation of learning a single rating per word and show that multi-sense abstractness ratings are potentially useful for metaphor detection. Finally, with this paper we publish automatically created abstractness norms for 3million English words and multi-words as well as automatically created sense specific abstractness ratings
Tasks Semantic Textual Similarity, Topic Models, Word Embeddings
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-1903/
PDF https://www.aclweb.org/anthology/W17-1903
PWC https://paperswithcode.com/paper/improving-verb-metaphor-detection-by
Repo
Framework

Rethinking Reprojection: Closing the Loop for Pose-Aware Shape Reconstruction From a Single Image

Title Rethinking Reprojection: Closing the Loop for Pose-Aware Shape Reconstruction From a Single Image
Authors Rui Zhu, Hamed Kiani Galoogahi, Chaoyang Wang, Simon Lucey
Abstract An emerging problem in computer vision is the reconstruction of 3D shape and pose of an object from a single image. Hitherto, the problem has been addressed through the application of canonical deep learning methods to regress from the image directly to the 3D shape and pose labels. These approaches, however, are problematic from two perspectives. First, they are minimizing the error between 3D shapes and pose labels - with little thought about the nature of this “label error” when reprojecting the shape back onto the image. Second, they rely on the onerous and ill-posed task of hand labeling natural images with respect to 3D shape and pose. In this paper we define the new task of pose-aware shape reconstruction from a single image, and we advocate that cheaper 2D annotations of objects silhouettes in natural images can be utilized. We design architectures of pose-aware shape reconstruction which reproject the predicted shape back on to the image using the predicted pose. Our evaluation on several object categories demonstrates the superiority of our method for predicting pose-aware 3D shapes from natural images.
Tasks
Published 2017-10-01
URL http://openaccess.thecvf.com/content_iccv_2017/html/Zhu_Rethinking_Reprojection_Closing_ICCV_2017_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2017/papers/Zhu_Rethinking_Reprojection_Closing_ICCV_2017_paper.pdf
PWC https://paperswithcode.com/paper/rethinking-reprojection-closing-the-loop-for-1
Repo
Framework

From which world is your graph

Title From which world is your graph
Authors Cheng Li, Felix Mf Wong, Zhenming Liu, Varun Kanade
Abstract Discovering statistical structure from links is a fundamental problem in the analysis of social networks. Choosing a misspecified model, or equivalently, an incorrect inference algorithm will result in an invalid analysis or even falsely uncover patterns that are in fact artifacts of the model. This work focuses on unifying two of the most widely used link-formation models: the stochastic block model (SBM) and the small world (or latent space) model (SWM). Integrating techniques from kernel learning, spectral graph theory, and nonlinear dimensionality reduction, we develop the first statistically sound polynomial-time algorithm to discover latent patterns in sparse graphs for both models. When the network comes from an SBM, the algorithm outputs a block structure. When it is from an SWM, the algorithm outputs estimates of each node’s latent position.
Tasks Dimensionality Reduction
Published 2017-12-01
URL http://papers.nips.cc/paper/6745-from-which-world-is-your-graph
PDF http://papers.nips.cc/paper/6745-from-which-world-is-your-graph.pdf
PWC https://paperswithcode.com/paper/from-which-world-is-your-graph-1
Repo
Framework

A Layered Language Model based Hybrid Approach to Automatic Full Diacritization of Arabic

Title A Layered Language Model based Hybrid Approach to Automatic Full Diacritization of Arabic
Authors Mohamed Al-Badrashiny, Abdelati Hawwari, Mona Diab
Abstract In this paper we present a system for automatic Arabic text diacritization using three levels of analysis granularity in a layered back off manner. We build and exploit diacritized language models (LM) for each of three different levels of granularity: surface form, morphologically segmented into prefix/stem/suffix, and character level. For each of the passes, we use Viterbi search to pick the most probable diacritization per word in the input. We start with the surface form LM, followed by the morphological level, then finally we leverage the character level LM. Our system outperforms all of the published systems evaluated against the same training and test data. It achieves a 10.87{%} WER for complete full diacritization including lexical and syntactic diacritization, and 3.0{%} WER for lexical diacritization, ignoring syntactic diacritization.
Tasks Arabic Text Diacritization, Language Modelling, Machine Translation, Morphological Analysis, Transliteration, Word Sense Disambiguation
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-1321/
PDF https://www.aclweb.org/anthology/W17-1321
PWC https://paperswithcode.com/paper/a-layered-language-model-based-hybrid
Repo
Framework

The Karlsruhe Institute of Technology Systems for the News Translation Task in WMT 2017

Title The Karlsruhe Institute of Technology Systems for the News Translation Task in WMT 2017
Authors Ngoc-Quan Pham, Jan Niehues, Thanh-Le Ha, Eunah Cho, Matthias Sperber, Alex Waibel, er
Abstract
Tasks Domain Adaptation, Machine Translation, Tokenization
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4736/
PDF https://www.aclweb.org/anthology/W17-4736
PWC https://paperswithcode.com/paper/the-karlsruhe-institute-of-technology-systems-2
Repo
Framework

Japanese Sentence Compression with a Large Training Dataset

Title Japanese Sentence Compression with a Large Training Dataset
Authors Shun Hasegawa, Yuta Kikuchi, Hiroya Takamura, Manabu Okumura
Abstract In English, high-quality sentence compression models by deleting words have been trained on automatically created large training datasets. We work on Japanese sentence compression by a similar approach. To create a large Japanese training dataset, a method of creating English training dataset is modified based on the characteristics of the Japanese language. The created dataset is used to train Japanese sentence compression models based on the recurrent neural network.
Tasks Sentence Compression
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-2044/
PDF https://www.aclweb.org/anthology/P17-2044
PWC https://paperswithcode.com/paper/japanese-sentence-compression-with-a-large
Repo
Framework

Learning local and global contexts using a convolutional recurrent network model for relation classification in biomedical text

Title Learning local and global contexts using a convolutional recurrent network model for relation classification in biomedical text
Authors Desh Raj, SUNIL SAHU, Ashish Anand
Abstract
Tasks Dependency Parsing, Feature Engineering, Morphological Analysis, Relation Classification, Relation Extraction
Published 2017-08-01
URL https://www.aclweb.org/anthology/papers/K17-1032/k17-1032
PDF https://www.aclweb.org/anthology/K17-1032
PWC https://paperswithcode.com/paper/learning-local-and-global-contexts-using-a
Repo
Framework

Incorporating Word Reordering Knowledge into Attention-based Neural Machine Translation

Title Incorporating Word Reordering Knowledge into Attention-based Neural Machine Translation
Authors Jinchao Zhang, Mingxuan Wang, Qun Liu, Jie Zhou
Abstract This paper proposes three distortion models to explicitly incorporate the word reordering knowledge into attention-based Neural Machine Translation (NMT) for further improving translation performance. Our proposed models enable attention mechanism to attend to source words regarding both the semantic requirement and the word reordering penalty. Experiments on Chinese-English translation show that the approaches can improve word alignment quality and achieve significant translation improvements over a basic attention-based NMT by large margins. Compared with previous works on identical corpora, our system achieves the state-of-the-art performance on translation quality.
Tasks Machine Translation, Word Alignment
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-1140/
PDF https://www.aclweb.org/anthology/P17-1140
PWC https://paperswithcode.com/paper/incorporating-word-reordering-knowledge-into
Repo
Framework

A tool for extracting sense-disambiguated example sentences through user feedback

Title A tool for extracting sense-disambiguated example sentences through user feedback
Authors Beto Boullosa, Richard Eckart de Castilho, Alex Geyken, er, Lothar Lemnitzer, Iryna Gurevych
Abstract This paper describes an application system aimed to help lexicographers in the extraction of example sentences for a given headword based on its different senses. The tool uses classification and clustering methods and incorporates user feedback to refine its results.
Tasks
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-3018/
PDF https://www.aclweb.org/anthology/E17-3018
PWC https://paperswithcode.com/paper/a-tool-for-extracting-sense-disambiguated
Repo
Framework
comments powered by Disqus