January 25, 2020

2660 words 13 mins read

Paper Group NANR 23

Paper Group NANR 23

Corpus of Multimodal Interaction for Collaborative Planning. Contextualized Role Interaction for Neural Machine Translation. Building Detail-Sensitive Semantic Segmentation Networks With Polynomial Pooling. Exploring Pre-trained Language Models for Event Extraction and Generation. UaiNets: From Unsupervised to Active Deep Anomaly Detection. Summary …

Corpus of Multimodal Interaction for Collaborative Planning

Title Corpus of Multimodal Interaction for Collaborative Planning
Authors Miltiadis Marios Katsakioris, Helen Hastie, Ioannis Konstas, Atanas Laskov
Abstract As autonomous systems become more commonplace, we need a way to easily and naturally communicate to them our goals and collaboratively come up with a plan on how to achieve these goals. To this end, we conducted a Wizard of Oz study to gather data and investigate the way operators would collaboratively make plans via a conversational {`}planning assistant{'} for remote autonomous systems. We present here a corpus of 22 dialogs from expert operators, which can be used to train such a system. Data analysis shows that multimodality is key to successful interaction, measured both quantitatively and qualitatively via user feedback. |
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-1601/
PDF https://www.aclweb.org/anthology/W19-1601
PWC https://paperswithcode.com/paper/corpus-of-multimodal-interaction-for
Repo
Framework

Contextualized Role Interaction for Neural Machine Translation

Title Contextualized Role Interaction for Neural Machine Translation
Authors Dirk Weissenborn, Douwe Kiela, Jason Weston, Kyunghyun Cho
Abstract Word inputs tend to be represented as single continuous vectors in deep neural networks. It is left to the subsequent layers of the network to extract relevant aspects of a word’s meaning based on the context in which it appears. In this paper, we investigate whether word representations can be improved by explicitly incorporating the idea of latent roles. That is, we propose a role interaction layer (RIL) that consists of context-dependent (latent) role assignments and role-specific transformations. We evaluate the RIL on machine translation using two language pairs (En-De and En-Fi) and three datasets of varying size. We find that the proposed mechanism improves translation quality over strong baselines with limited amounts of data, but that the improvement diminishes as the size of data grows, indicating that powerful neural MT systems are capable of implicitly modeling role-word interaction by themselves. Our qualitative analysis reveals that the RIL extracts meaningful context-dependent roles and that it allows us to inspect more deeply the internal mechanisms of state-of-the-art neural machine translation systems.
Tasks Machine Translation
Published 2019-01-01
URL https://openreview.net/forum?id=ryx3_iAcY7
PDF https://openreview.net/pdf?id=ryx3_iAcY7
PWC https://paperswithcode.com/paper/contextualized-role-interaction-for-neural
Repo
Framework

Building Detail-Sensitive Semantic Segmentation Networks With Polynomial Pooling

Title Building Detail-Sensitive Semantic Segmentation Networks With Polynomial Pooling
Authors Zhen Wei, Jingyi Zhang, Li Liu, Fan Zhu, Fumin Shen, Yi Zhou, Si Liu, Yao Sun, Ling Shao
Abstract Semantic segmentation is an important computer vision task, which aims to allocate a semantic label to each pixel in an image. When training a segmentation model, it is common to fine-tune a classification network pre-trained on a large-scale dataset. However, as an intrinsic property of the classification model, invariance to spatial perturbation resulting from the lose of detail-sensitivity prevents segmentation networks from achieving high performance. The use of standard poolings is one of the key factors for this invariance. The most common standard poolings are max and average pooling. Max pooling can increase both the invariance to spatial perturbations and the non-linearity of the networks. Average pooling, on the other hand, is sensitive to spatial perturbations, but is a linear function. For semantic segmentation, we prefer both the preservation of detailed cues within a local feature region and non-linearity that increases a network’s functional complexity. In this work, we propose a polynomial pooling (P-pooling) function that finds an intermediate form between max and average pooling to provide an optimally balanced and self-adjusted pooling strategy for semantic segmentation. The P-pooling is differentiable and can be applied into a variety of pre-trained networks. Extensive studies on the PASCAL VOC, Cityscapes and ADE20k datasets demonstrate the superiority of P-pooling over other poolings. Experiments on various network architectures and state-of-the-art training strategies also show that models with P-pooling layers consistently outperform those directly fine-tuned using pre-trained classification models.
Tasks Semantic Segmentation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Wei_Building_Detail-Sensitive_Semantic_Segmentation_Networks_With_Polynomial_Pooling_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Wei_Building_Detail-Sensitive_Semantic_Segmentation_Networks_With_Polynomial_Pooling_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/building-detail-sensitive-semantic
Repo
Framework

Exploring Pre-trained Language Models for Event Extraction and Generation

Title Exploring Pre-trained Language Models for Event Extraction and Generation
Authors Sen Yang, Dawei Feng, Linbo Qiao, Zhigang Kan, Dongsheng Li
Abstract Traditional approaches to the task of ACE event extraction usually depend on manually annotated data, which is often laborious to create and limited in size. Therefore, in addition to the difficulty of event extraction itself, insufficient training data hinders the learning process as well. To promote event extraction, we first propose an event extraction model to overcome the roles overlap problem by separating the argument prediction in terms of roles. Moreover, to address the problem of insufficient training data, we propose a method to automatically generate labeled data by editing prototypes and screen out generated samples by ranking the quality. Experiments on the ACE2005 dataset demonstrate that our extraction model can surpass most existing extraction methods. Besides, incorporating our generation method exhibits further significant improvement. It obtains new state-of-the-art results on the event extraction task, including pushing the F1 score of trigger classification to 81.1{%}, and the F1 score of argument classification to 58.9{%}.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1522/
PDF https://www.aclweb.org/anthology/P19-1522
PWC https://paperswithcode.com/paper/exploring-pre-trained-language-models-for
Repo
Framework

UaiNets: From Unsupervised to Active Deep Anomaly Detection

Title UaiNets: From Unsupervised to Active Deep Anomaly Detection
Authors Tiago Pimentel, Marianne Monteiro, Juliano Viana, Adriano Veloso, Nivio Ziviani
Abstract This work presents a method for active anomaly detection which can be built upon existing deep learning solutions for unsupervised anomaly detection. We show that a prior needs to be assumed on what the anomalies are, in order to have performance guarantees in unsupervised anomaly detection. We argue that active anomaly detection has, in practice, the same cost of unsupervised anomaly detection but with the possibility of much better results. To solve this problem, we present a new layer that can be attached to any deep learning model designed for unsupervised anomaly detection to transform it into an active method, presenting results on both synthetic and real anomaly detection datasets.
Tasks Anomaly Detection, Unsupervised Anomaly Detection
Published 2019-05-01
URL https://openreview.net/forum?id=HJex0o05F7
PDF https://openreview.net/pdf?id=HJex0o05F7
PWC https://paperswithcode.com/paper/uainets-from-unsupervised-to-active-deep
Repo
Framework

Summary Refinement through Denoising

Title Summary Refinement through Denoising
Authors Nikola Nikolov, Aless Calmanovici, ro, Richard Hahnloser
Abstract We propose a simple method for post-processing the outputs of a text summarization system in order to refine its overall quality. Our approach is to train text-to-text rewriting models to correct information redundancy errors that may arise during summarization. We train on synthetically generated noisy summaries, testing three different types of noise that introduce out-of-context information within each summary. When applied on top of extractive and abstractive summarization baselines, our summary denoising models yield metric improvements while reducing redundancy.
Tasks Abstractive Text Summarization, Denoising, Text Summarization
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-1097/
PDF https://www.aclweb.org/anthology/R19-1097
PWC https://paperswithcode.com/paper/summary-refinement-through-denoising-1
Repo
Framework

Augmenting Abstract Meaning Representation for Human-Robot Dialogue

Title Augmenting Abstract Meaning Representation for Human-Robot Dialogue
Authors Claire Bonial, Lucia Donatelli, Stephanie M. Lukin, Stephen Tratz, Ron Artstein, David Traum, Clare Voss
Abstract We detail refinements made to Abstract Meaning Representation (AMR) that make the representation more suitable for supporting a situated dialogue system, where a human remotely controls a robot for purposes of search and rescue and reconnaissance. We propose 36 augmented AMRs that capture speech acts, tense and aspect, and spatial information. This linguistic information is vital for representing important distinctions, for example whether the robot has moved, is moving, or will move. We evaluate two existing AMR parsers for their performance on dialogue data. We also outline a model for graph-to-graph conversion, in which output from AMR parsers is converted into our refined AMRs. The design scheme presented here, though task-specific, is extendable for broad coverage of speech acts using AMR in future task-independent work.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3322/
PDF https://www.aclweb.org/anthology/W19-3322
PWC https://paperswithcode.com/paper/augmenting-abstract-meaning-representation
Repo
Framework

Entangled Transformer for Image Captioning

Title Entangled Transformer for Image Captioning
Authors Guang Li, Linchao Zhu, Ping Liu, Yi Yang
Abstract In image captioning, the typical attention mechanisms are arduous to identify the equivalent visual signals especially when predicting highly abstract words. This phenomenon is known as the semantic gap between vision and language. This problem can be overcome by providing semantic attributes that are homologous to language. Thanks to the inherent recurrent nature and gated operating mechanism, Recurrent Neural Network (RNN) and its variants are the dominating architectures in image captioning. However, when designing elaborate attention mechanisms to integrate visual inputs and semantic attributes, RNN-like variants become unflexible due to their complexities. In this paper, we investigate a Transformer-based sequence modeling framework, built only with attention layers and feedforward layers. To bridge the semantic gap, we introduce EnTangled Attention (ETA) that enables the Transformer to exploit semantic and visual information simultaneously. Furthermore, Gated Bilateral Controller (GBC) is proposed to guide the interactions between the multimodal information. We name our model as ETA-Transformer. Remarkably, ETA-Transformer achieves state-of-the-art performance on the MSCOCO image captioning dataset. The ablation studies validate the improvements of our proposed modules.
Tasks Image Captioning
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Li_Entangled_Transformer_for_Image_Captioning_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Li_Entangled_Transformer_for_Image_Captioning_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/entangled-transformer-for-image-captioning
Repo
Framework

How Important is a Neuron

Title How Important is a Neuron
Authors Kedar Dhamdhere, Mukund Sundararajan, Qiqi Yan
Abstract The problem of attributing a deep network’s prediction to its input/base features is well-studied (cf. Simonyan et al. (2013)). We introduce the notion of conductance to extend the notion of attribution to understanding the importance of hidden units. Informally, the conductance of a hidden unit of a deep network is the flow of attribution via this hidden unit. We can use conductance to understand the importance of a hidden unit to the prediction for a specific input, or over a set of inputs. We justify conductance in multiple ways via a qualitative comparison with other methods, via some axiomatic results, and via an empirical evaluation based on a feature selection task. The empirical evaluations are done using the Inception network over ImageNet data, and a convolutinal network over text data. In both cases, we demonstrate the effectiveness of conductance in identifying interesting insights about the internal workings of these networks.
Tasks Feature Selection
Published 2019-05-01
URL https://openreview.net/forum?id=SylKoo0cKm
PDF https://openreview.net/pdf?id=SylKoo0cKm
PWC https://paperswithcode.com/paper/how-important-is-a-neuron-1
Repo
Framework

Give It a Shot: Few-shot Learning to Normalize ADR Mentions in Social Media Posts

Title Give It a Shot: Few-shot Learning to Normalize ADR Mentions in Social Media Posts
Authors Emmanouil Manousogiannis, Sepideh Mesbah, Aless Bozzon, ro, Selene Baez, Robert Jan Sips
Abstract This paper describes the system that team MYTOMORROWS-TU DELFT developed for the 2019 Social Media Mining for Health Applications (SMM4H) Shared Task 3, for the end-to-end normalization of ADR tweet mentions to their corresponding MEDDRA codes. For the first two steps, we reuse a state-of-the art approach, focusing our contribution on the final entity-linking step. For that we propose a simple Few-Shot learning approach, based on pre-trained word embeddings and data from the UMLS, combined with the provided training data. Our system (relaxed F1: 0.337-0.345) outperforms the average (relaxed F1 0.2972) of the participants in this task, demonstrating the potential feasibility of few-shot learning in the context of medical text normalization.
Tasks Entity Linking, Few-Shot Learning, Word Embeddings
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3219/
PDF https://www.aclweb.org/anthology/W19-3219
PWC https://paperswithcode.com/paper/give-it-a-shot-few-shot-learning-to-normalize
Repo
Framework

Perception-Aware Point-Based Value Iteration for Partially Observable Markov Decision Processes

Title Perception-Aware Point-Based Value Iteration for Partially Observable Markov Decision Processes
Authors Mahsa Ghasemi, Ufuk Topcu
Abstract Partially observable Markov decision processes (POMDPs) are a widely-used framework to model decision-making with uncertainty about the environment and under stochastic outcome. In conventional POMDP models, the observations that the agent receives originate from fixed known distribution. However, in a variety of real-world scenarios the agent has an active role in its perception by selecting which observations to receive. Due to combinatorial nature of such selection process, it is computationally intractable to integrate the perception decision with the planning decision. To prevent such expansion of the action space, we propose a greedy strategy for observation selection that aims to minimize the uncertainty in state. We develop a novel point-based value iteration algorithm that incorporates the greedy strategy to achieve near-optimal uncertainty reduction for sampled belief points. This in turn enables the solver to efficiently approximate the reachable subspace of belief simplex by essentially separating computations related to perception from planning. Lastly, we implement the proposed solver and demonstrate its performance and computational advantage in a range of robotic scenarios where the robot simultaneously performs active perception and planning.
Tasks Decision Making
Published 2019-05-01
URL https://openreview.net/forum?id=S1lTg3RcFm
PDF https://openreview.net/pdf?id=S1lTg3RcFm
PWC https://paperswithcode.com/paper/perception-aware-point-based-value-iteration
Repo
Framework

Revisiting the Role of Feature Engineering for Compound Type Identification in Sanskrit

Title Revisiting the Role of Feature Engineering for Compound Type Identification in Sanskrit
Authors S, Jivnesh han, Amrith Krishna, Pawan Goyal, Laxmidhar Behera
Abstract
Tasks Feature Engineering
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-7503/
PDF https://www.aclweb.org/anthology/W19-7503
PWC https://paperswithcode.com/paper/revisiting-the-role-of-feature-engineering
Repo
Framework

Cross-Lingual Word Embeddings and the Structure of the Human Bilingual Lexicon

Title Cross-Lingual Word Embeddings and the Structure of the Human Bilingual Lexicon
Authors Paola Merlo, Maria Andueza Rodriguez
Abstract Research on the bilingual lexicon has uncovered fascinating interactions between the lexicons of the native language and of the second language in bilingual speakers. In particular, it has been found that the lexicon of the underlying native language affects the organisation of the second language. In the spirit of interpreting current distributed representations, this paper investigates two models of cross-lingual word embeddings, comparing them to the shared-translation effect and the cross-lingual coactivation effects of false and true friends (cognates) found in humans. We find that the similarity structure of the cross-lingual word embeddings space yields the same effects as the human bilingual lexicon.
Tasks Word Embeddings
Published 2019-11-01
URL https://www.aclweb.org/anthology/K19-1011/
PDF https://www.aclweb.org/anthology/K19-1011
PWC https://paperswithcode.com/paper/cross-lingual-word-embeddings-and-the
Repo
Framework

Learning Conceptual Spaces with Disentangled Facets

Title Learning Conceptual Spaces with Disentangled Facets
Authors Rana Alshaikh, Zied Bouraoui, Steven Schockaert
Abstract Conceptual spaces are geometric representations of meaning that were proposed by G ̈ardenfors (2000). They share many similarities with the vector space embeddings that are commonly used in natural language processing. However, rather than representing entities in a single vector space, conceptual spaces are usually decomposed into several facets, each of which is then modelled as a relatively low dimensional vector space. Unfortunately, the problem of learning such conceptual spaces has thus far only received limited attention. To address this gap, we analyze how, and to what extent, a given vector space embedding can be decomposed into meaningful facets in an unsupervised fashion. While this problem is highly challenging, we show that useful facets can be discovered by relying on word embeddings to group semantically related features.
Tasks Word Embeddings
Published 2019-11-01
URL https://www.aclweb.org/anthology/K19-1013/
PDF https://www.aclweb.org/anthology/K19-1013
PWC https://paperswithcode.com/paper/learning-conceptual-spaces-with-disentangled
Repo
Framework

Cost-Sensitive BERT for Generalisable Sentence Classification on Imbalanced Data

Title Cost-Sensitive BERT for Generalisable Sentence Classification on Imbalanced Data
Authors Harish Tayyar Madabushi, Elena Kochkina, Michael Castelle
Abstract The automatic identification of propaganda has gained significance in recent years due to technological and social changes in the way news is generated and consumed. That this task can be addressed effectively using BERT, a powerful new architecture which can be fine-tuned for text classification tasks, is not surprising. However, propaganda detection, like other tasks that deal with news documents and other forms of decontextualized social communication (e.g. sentiment analysis), inherently deals with data whose categories are simultaneously imbalanced and dissimilar. We show that BERT, while capable of handling imbalanced classes with no additional data augmentation, does not generalise well when the training and test data are sufficiently dissimilar (as is often the case with news sources, whose topics evolve over time). We show how to address this problem by providing a statistical measure of similarity between datasets and a method of incorporating cost-weighting into BERT when the training and test sets are dissimilar. We test these methods on the Propaganda Techniques Corpus (PTC) and achieve the second highest score on sentence-level propaganda classification.
Tasks Data Augmentation, Sentence Classification, Sentiment Analysis, Text Classification
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5018/
PDF https://www.aclweb.org/anthology/D19-5018
PWC https://paperswithcode.com/paper/cost-sensitive-bert-for-generalisable
Repo
Framework
comments powered by Disqus