January 24, 2020

2586 words 13 mins read

Paper Group NANR 217

Paper Group NANR 217

CDTB: A Color and Depth Visual Object Tracking Dataset and Benchmark. Read, Attend and Comment: A Deep Architecture for Automatic NewsComment Generation. An Encoder with non-Sequential Dependency for Neural Data-to-Text Generation. A Closer Look at Recent Results of Verb Selection for Data-to-Text NLG. Detecting Out-Of-Distribution Samples Using Lo …

CDTB: A Color and Depth Visual Object Tracking Dataset and Benchmark

Title CDTB: A Color and Depth Visual Object Tracking Dataset and Benchmark
Authors Alan Lukezic, Ugur Kart, Jani Kapyla, Ahmed Durmush, Joni-Kristian Kamarainen, Jiri Matas, Matej Kristan
Abstract We propose a new color-and-depth general visual object tracking benchmark (CDTB). CDTB is recorded by several passive and active RGB-D setups and contains indoor as well as outdoor sequences acquired in direct sunlight. The CDTB dataset is the largest and most diverse dataset in RGB-D tracking, with an order of magnitude larger number of frames than related datasets. The sequences have been carefully recorded to contain significant object pose change, clutter, occlusion, and periods of long-term target absence to enable tracker evaluation under realistic conditions. Sequences are per-frame annotated with 13 visual attributes for detailed analysis. Experiments with RGB and RGB-D trackers show that CDTB is more challenging than previous datasets. State-of-the-art RGB trackers outperform the recent RGB-D trackers, indicating a large gap between the two fields, which has not been previously detected by the prior benchmarks. Based on the results of the analysis we point out opportunities for future research in RGB-D tracker design.
Tasks Object Tracking, Visual Object Tracking
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Lukezic_CDTB_A_Color_and_Depth_Visual_Object_Tracking_Dataset_and_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Lukezic_CDTB_A_Color_and_Depth_Visual_Object_Tracking_Dataset_and_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/cdtb-a-color-and-depth-visual-object-tracking-1
Repo
Framework

Read, Attend and Comment: A Deep Architecture for Automatic NewsComment Generation

Title Read, Attend and Comment: A Deep Architecture for Automatic NewsComment Generation
Authors Ze Yang, Can Xu, Wei Wu, Zhoujun Li
Abstract Automatic news comment generation is a new testbed for techniques of natural language generation. In this paper, we propose a “read-attend-comment” procedure for news comment generation and formalize the procedure with a reading network and a generation network. The reading network comprehends a news article and distills some important points from it, then the generation network creates a comment by attending to the extracted discrete points and the news title. We optimize the model in an end-to-end manner by maximizing a variational lower bound of the true objective using the back-propagation algorithm. Experimental results on two datasets indicate that our model can significantly outperform existing methods in terms of both automatic evaluation and human judgment.
Tasks Text Generation
Published 2019-10-01
URL https://arxiv.org/abs/1909.11974
PDF https://arxiv.org/pdf/1909.11974.pdf
PWC https://paperswithcode.com/paper/read-attend-and-comment-a-deep-architecture-1
Repo
Framework

An Encoder with non-Sequential Dependency for Neural Data-to-Text Generation

Title An Encoder with non-Sequential Dependency for Neural Data-to-Text Generation
Authors Feng Nie, Jinpeng Wang, Rong Pan, Chin-Yew Lin
Abstract Data-to-text generation aims to generate descriptions given a structured input data (i.e., a table with multiple records). Existing neural methods for encoding input data can be divided into two categories: a) pooling based encoders which ignore dependencies between input records or b) recurrent encoders which model only sequential dependencies between input records. In our investigation, although the recurrent encoder generally outperforms the pooling based encoder by learning the sequential dependencies, it is sensitive to the order of the input records (i.e., performance decreases when injecting the random shuffling noise over input data). To overcome this problem, we propose to adopt the self-attention mechanism to learn dependencies between arbitrary input records. Experimental results show the proposed method achieves comparable results and remains stable under random shuffling over input data.
Tasks Data-to-Text Generation, Text Generation
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-8619/
PDF https://www.aclweb.org/anthology/W19-8619
PWC https://paperswithcode.com/paper/an-encoder-with-non-sequential-dependency-for
Repo
Framework

A Closer Look at Recent Results of Verb Selection for Data-to-Text NLG

Title A Closer Look at Recent Results of Verb Selection for Data-to-Text NLG
Authors Guanyi Chen, Jin-Ge Yao
Abstract Automatic natural language generation systems need to use the contextually-appropriate verbs when describing different kinds of facts or events, which has triggered research interest on verb selection for data-to-text generation. In this paper, we discuss a few limitations of the current task settings and the evaluation metrics. We also provide two simple, efficient, interpretable baseline approaches for statistical selection of trend verbs, which give a strong performance on both previously used evaluation metrics and our new evaluation.
Tasks Data-to-Text Generation, Text Generation
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-8622/
PDF https://www.aclweb.org/anthology/W19-8622
PWC https://paperswithcode.com/paper/a-closer-look-at-recent-results-of-verb
Repo
Framework

Detecting Out-Of-Distribution Samples Using Low-Order Deep Features Statistics

Title Detecting Out-Of-Distribution Samples Using Low-Order Deep Features Statistics
Authors Igor M. Quintanilha, Roberto de M. E. Filho, José Lezama, Mauricio Delbracio, Leonardo O. Nunes
Abstract The ability to detect when an input sample was not drawn from the training distribution is an important desirable property of deep neural networks. In this paper, we show that a simple ensembling of first and second order deep feature statistics can be exploited to effectively differentiate in-distribution and out-of-distribution samples. Specifically, we observe that the mean and standard deviation within feature maps differs greatly between in-distribution and out-of-distribution samples. Based on this observation, we propose a simple and efficient plug-and-play detection procedure that does not require re-training, pre-processing or changes to the model. The proposed method outperforms the state-of-the-art by a large margin in all standard benchmarking tasks, while being much simpler to implement and execute. Notably, our method improves the true negative rate from 39.6% to 95.3% when 95% of in-distribution (CIFAR-100) are correctly detected using a DenseNet and the out-of-distribution dataset is TinyImageNet resize. The source code of our method will be made publicly available.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=rkgpCoRctm
PDF https://openreview.net/pdf?id=rkgpCoRctm
PWC https://paperswithcode.com/paper/detecting-out-of-distribution-samples-using
Repo
Framework

On the effect of the activation function on the distribution of hidden nodes in a deep network

Title On the effect of the activation function on the distribution of hidden nodes in a deep network
Authors Philip M. Long and Hanie Sedghi
Abstract We analyze the joint probability distribution on the lengths of the vectors of hidden variables in different layers of a fully connected deep network, when the weights and biases are chosen randomly according to Gaussian distributions, and the input is binary-valued. We show that, if the activation function satisfies a minimal set of assumptions, satisfied by all activation functions that we know that are used in practice, then, as the width of the network gets large, the ``length process’’ converges in probability to a length map that is determined as a simple function of the variances of the random weights and biases, and the activation function. We also show that this convergence may fail for activation functions that violate our assumptions. |
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=HJej3s09Km
PDF https://openreview.net/pdf?id=HJej3s09Km
PWC https://paperswithcode.com/paper/on-the-effect-of-the-activation-function-on-1
Repo
Framework

Neural News Recommendation with Long- and Short-term User Representations

Title Neural News Recommendation with Long- and Short-term User Representations
Authors Mingxiao An, Fangzhao Wu, Chuhan Wu, Kun Zhang, Zheng Liu, Xing Xie
Abstract Personalized news recommendation is important to help users find their interested news and improve reading experience. A key problem in news recommendation is learning accurate user representations to capture their interests. Users usually have both long-term preferences and short-term interests. However, existing news recommendation methods usually learn single representations of users, which may be insufficient. In this paper, we propose a neural news recommendation approach which can learn both long- and short-term user representations. The core of our approach is a news encoder and a user encoder. In the news encoder, we learn representations of news from their titles and topic categories, and use attention network to select important words. In the user encoder, we propose to learn long-term user representations from the embeddings of their IDs.In addition, we propose to learn short-term user representations from their recently browsed news via GRU network. Besides, we propose two methods to combine long-term and short-term user representations. The first one is using the long-term user representation to initialize the hidden state of the GRU network in short-term user representation. The second one is concatenating both long- and short-term user representations as a unified user vector. Extensive experiments on a real-world dataset show our approach can effectively improve the performance of neural news recommendation.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1033/
PDF https://www.aclweb.org/anthology/P19-1033
PWC https://paperswithcode.com/paper/neural-news-recommendation-with-long-and
Repo
Framework

Making Sense of Conflicting (Defeasible) Rules in the Controlled Natural Language ACE: Design of a System with Support for Existential Quantification Using Skolemization

Title Making Sense of Conflicting (Defeasible) Rules in the Controlled Natural Language ACE: Design of a System with Support for Existential Quantification Using Skolemization
Authors Martin Diller, Adam Wyner, Hannes Strass
Abstract We present the design of a system for making sense of conflicting rules expressed in a fragment of the prominent controlled natural language ACE, yet extended with means of expressing defeasible rules in the form of normality assumptions. The approach we describe is ultimately based on answer-set-programming (ASP); simulating existential quantification by using skolemization in a manner resembling a translation for ASP recently formalized in the context of ∃-ASP. We discuss the advantages of this approach to building on the existing ACE interface to rule-systems, ACERules.
Tasks
Published 2019-05-01
URL https://www.aclweb.org/anthology/W19-0505/
PDF https://www.aclweb.org/anthology/W19-0505
PWC https://paperswithcode.com/paper/making-sense-of-conflicting-defeasible-rules
Repo
Framework

Annotation and Automatic Classification of Aspectual Categories

Title Annotation and Automatic Classification of Aspectual Categories
Authors Markus Egg, Helena Prepens, Will Roberts
Abstract We present the first annotated resource for the aspectual classification of German verb tokens in their clausal context. We use aspectual features compatible with the plurality of aspectual classifications in previous work and treat aspectual ambiguity systematically. We evaluate our corpus by using it to train supervised classifiers to automatically assign aspectual categories to verbs in context, permitting favourable comparisons to previous work.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1323/
PDF https://www.aclweb.org/anthology/P19-1323
PWC https://paperswithcode.com/paper/annotation-and-automatic-classification-of
Repo
Framework

Filling Gender & Number Gaps in Neural Machine Translation with Black-box Context Injection

Title Filling Gender & Number Gaps in Neural Machine Translation with Black-box Context Injection
Authors Amit Moryossef, Roee Aharoni, Yoav Goldberg
Abstract When translating from a language that does not morphologically mark information such as gender and number into a language that does, translation systems must {``}guess{''} this missing information, often leading to incorrect translations in the given context. We propose a black-box approach for injecting the missing information to a pre-trained neural machine translation system, allowing to control the morphological variations in the generated translations without changing the underlying model or training data. We evaluate our method on an English to Hebrew translation task, and show that it is effective in injecting the gender and number information and that supplying the correct information improves the translation accuracy in up to 2.3 BLEU on a female-speaker test set for a state-of-the-art online black-box system. Finally, we perform a fine-grained syntactic analysis of the generated translations that shows the effectiveness of our method. |
Tasks Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3807/
PDF https://www.aclweb.org/anthology/W19-3807
PWC https://paperswithcode.com/paper/filling-gender-number-gaps-in-neural-machine-1
Repo
Framework

English-Indonesian Neural Machine Translation for Spoken Language Domains

Title English-Indonesian Neural Machine Translation for Spoken Language Domains
Authors Meisyarah Dwiastuti
Abstract In this work, we conduct a study on Neural Machine Translation (NMT) for English-Indonesian (EN-ID) and Indonesian-English (ID-EN). We focus on spoken language domains, namely colloquial and speech languages. We build NMT systems using the Transformer model for both translation directions and implement domain adaptation, in which we train our pre-trained NMT systems on speech language (in-domain) data. Moreover, we conduct an evaluation on how the domain-adaptation method in our EN-ID system can result in more formal translation outputs.
Tasks Domain Adaptation, Machine Translation
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-2043/
PDF https://www.aclweb.org/anthology/P19-2043
PWC https://paperswithcode.com/paper/english-indonesian-neural-machine-translation
Repo
Framework

Knowledge Discovery and Hypothesis Generation from Online Patient Forums: A Research Proposal

Title Knowledge Discovery and Hypothesis Generation from Online Patient Forums: A Research Proposal
Authors Anne Dirkson
Abstract The unprompted patient experiences shared on patient forums contain a wealth of unexploited knowledge. Mining this knowledge and cross-linking it with biomedical literature, could expose novel insights, which could subsequently provide hypotheses for further clinical research. As of yet, automated methods for open knowledge discovery on patient forum text are lacking. Thus, in this research proposal, we outline future research into methods for mining, aggregating and cross-linking patient knowledge from online forums. Additionally, we aim to address how one could measure the credibility of this extracted knowledge.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-2009/
PDF https://www.aclweb.org/anthology/P19-2009
PWC https://paperswithcode.com/paper/knowledge-discovery-and-hypothesis-generation
Repo
Framework

Cross-lingual Joint Entity and Word Embedding to Improve Entity Linking and Parallel Sentence Mining

Title Cross-lingual Joint Entity and Word Embedding to Improve Entity Linking and Parallel Sentence Mining
Authors Xiaoman Pan, Thamme Gowda, Heng Ji, Jonathan May, Scott Miller
Abstract Entities, which refer to distinct objects in the real world, can be viewed as language universals and used as effective signals to generate less ambiguous semantic representations and align multiple languages. We propose a novel method, CLEW, to generate cross-lingual data that is a mix of entities and contextual words based on Wikipedia. We replace each anchor link in the source language with its corresponding entity title in the target language if it exists, or in the source language otherwise. A cross-lingual joint entity and word embedding learned from this kind of data not only can disambiguate linkable entities but can also effectively represent unlinkable entities. Because this multilingual common space directly relates the semantics of contextual words in the source language to that of entities in the target language, we leverage it for unsupervised cross-lingual entity linking. Experimental results show that CLEW significantly advances the state-of-the-art: up to 3.1{%} absolute F-score gain for unsupervised cross-lingual entity linking. Moreover, it provides reliable alignment on both the word/entity level and the sentence level, and thus we use it to mine parallel sentences for all (302, 2) language pairs in Wikipedia.
Tasks Cross-Lingual Entity Linking, Entity Linking
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-6107/
PDF https://www.aclweb.org/anthology/D19-6107
PWC https://paperswithcode.com/paper/cross-lingual-joint-entity-and-word-embedding
Repo
Framework

Speaker-adapted neural-network-based fusion for multimodal reference resolution

Title Speaker-adapted neural-network-based fusion for multimodal reference resolution
Authors Diana Kleingarn, Nima Nabizadeh, Martin Heckmann, Dorothea Kolossa
Abstract Humans use a variety of approaches to reference objects in the external world, including verbal descriptions, hand and head gestures, eye gaze or any combination of them. The amount of useful information from each modality, however, may vary depending on the specific person and on several other factors. For this reason, it is important to learn the correct combination of inputs for inferring the best-fitting reference. In this paper, we investigate appropriate speaker-dependent and independent fusion strategies in a multimodal reference resolution task. We show that without any change in the modality models, only through an optimized fusion technique, it is possible to reduce the error rate of the system on a reference resolution task by more than 50{%}.
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-5925/
PDF https://www.aclweb.org/anthology/W19-5925
PWC https://paperswithcode.com/paper/speaker-adapted-neural-network-based-fusion
Repo
Framework

Honkling: In-Browser Personalization for Ubiquitous Keyword Spotting

Title Honkling: In-Browser Personalization for Ubiquitous Keyword Spotting
Authors Jaejun Lee, Raphael Tang, Jimmy Lin
Abstract Used for simple commands recognition on devices from smart speakers to mobile phones, keyword spotting systems are everywhere. Ubiquitous as well are web applications, which have grown in popularity and complexity over the last decade. However, despite their obvious advantages in natural language interaction, voice-enabled web applications are still few and far between. We attempt to bridge this gap with Honkling, a novel, JavaScript-based keyword spotting system. Purely client-side and cross-device compatible, Honkling can be deployed directly on user devices. Our in-browser implementation enables seamless personalization, which can greatly improve model quality; in the presence of underrepresented, non-American user accents, we can achieve up to an absolute 10{%} increase in accuracy in the personalized model with only a few examples.
Tasks Keyword Spotting
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-3016/
PDF https://www.aclweb.org/anthology/D19-3016
PWC https://paperswithcode.com/paper/honkling-in-browser-personalization-for
Repo
Framework
comments powered by Disqus