January 25, 2020

2396 words 12 mins read

Paper Group NANR 81

Paper Group NANR 81

A Large-Scale User Study of an Alexa Prize Chatbot: Effect of TTS Dynamism on Perceived Quality of Social Dialog. Robust Subspace Clustering With Independent and Piecewise Identically Distributed Noise Modeling. ParHistVis: Visualization of Parallel Multilingual Historical Data. WMDO: Fluency-based Word Mover’s Distance for Machine Translation Eval …

A Large-Scale User Study of an Alexa Prize Chatbot: Effect of TTS Dynamism on Perceived Quality of Social Dialog

Title A Large-Scale User Study of an Alexa Prize Chatbot: Effect of TTS Dynamism on Perceived Quality of Social Dialog
Authors Michelle Cohn, Chun-Yen Chen, Zhou Yu
Abstract This study tests the effect of cognitive-emotional expression in an Alexa text-to-speech (TTS) voice on users{'} experience with a social dialog system. We systematically introduced emotionally expressive interjections (e.g., {}Wow!{''}) and filler words (e.g., {}um{''}, {``}mhmm{''}) in an Amazon Alexa Prize socialbot, Gunrock. We tested whether these TTS manipulations improved users{'} ratings of their conversation across thousands of real user interactions (n=5,527). Results showed that interjections and fillers each improved users{'} holistic ratings, an improvement that further increased if the system used both manipulations. A separate perception experiment corroborated the findings from the user study, with improved social ratings for conversations including interjections; however, no positive effect was observed for fillers, suggesting that the role of the rater in the conversation{—}as active participant or external listener{—}is an important factor in assessing social dialogs. |
Tasks Chatbot
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-5935/
PDF https://www.aclweb.org/anthology/W19-5935
PWC https://paperswithcode.com/paper/a-large-scale-user-study-of-an-alexa-prize
Repo
Framework

Robust Subspace Clustering With Independent and Piecewise Identically Distributed Noise Modeling

Title Robust Subspace Clustering With Independent and Piecewise Identically Distributed Noise Modeling
Authors Yuanman Li, Jiantao Zhou, Xianwei Zheng, Jinyu Tian, Yuan Yan Tang
Abstract Most of the existing subspace clustering (SC) frameworks assume that the noise contaminating the data is generated by an independent and identically distributed (i.i.d.) source, where the Gaussianity is often imposed. Though these assumptions greatly simplify the underlying problems, they do not hold in many real-world applications. For instance, in face clustering, the noise is usually caused by random occlusions, local variations and unconstrained illuminations, which is essentially structural and hence satisfies neither the i.i.d. property nor the Gaussianity. In this work, we propose an independent and piecewise identically distributed (i.p.i.d.) noise model, where the i.i.d. property only holds locally. We demonstrate that the i.p.i.d. model better characterizes the noise encountered in practical scenarios, and accommodates the traditional i.i.d. model as a special case. Assisted by this generalized noise model, we design an information theoretic learning (ITL) framework for robust SC through a novel minimum weighted error entropy (MWEE) criterion. Extensive experimental results show that our proposed SC scheme significantly outperforms the state-of-the-art competing algorithms.
Tasks
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Li_Robust_Subspace_Clustering_With_Independent_and_Piecewise_Identically_Distributed_Noise_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Li_Robust_Subspace_Clustering_With_Independent_and_Piecewise_Identically_Distributed_Noise_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/robust-subspace-clustering-with-independent
Repo
Framework

ParHistVis: Visualization of Parallel Multilingual Historical Data

Title ParHistVis: Visualization of Parallel Multilingual Historical Data
Authors Aikaterini-Lida Kalouli, Rebecca Kehlbeck, Rita Sevastjanova, Katharina Kaiser, Georg A. Kaiser, Miriam Butt
Abstract The study of language change through parallel corpora can be advantageous for the analysis of complex interactions between time, text domain and language. Often, those advantages cannot be fully exploited due to the sparse but high-dimensional nature of such historical data. To tackle this challenge, we introduce ParHistVis: a novel, free, easy-to-use, interactive visualization tool for parallel, multilingual, diachronic and synchronic linguistic data. We illustrate the suitability of the components of the tool based on a use case of word order change in Romance wh-interrogatives.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4714/
PDF https://www.aclweb.org/anthology/W19-4714
PWC https://paperswithcode.com/paper/parhistvis-visualization-of-parallel
Repo
Framework

WMDO: Fluency-based Word Mover’s Distance for Machine Translation Evaluation

Title WMDO: Fluency-based Word Mover’s Distance for Machine Translation Evaluation
Authors Julian Chow, Lucia Specia, Pranava Madhyastha
Abstract We propose WMDO, a metric based on distance between distributions in the semantic vector space. Matching in the semantic space has been investigated for translation evaluation, but the constraints of a translation{'}s word order have not been fully explored. Building on the Word Mover{'}s Distance metric and various word embeddings, we introduce a fragmentation penalty to account for fluency of a translation. This word order extension is shown to perform better than standard WMD, with promising results against other types of metrics.
Tasks Machine Translation, Word Embeddings
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5356/
PDF https://www.aclweb.org/anthology/W19-5356
PWC https://paperswithcode.com/paper/wmdo-fluency-based-word-movers-distance-for
Repo
Framework

Don’t let your Discriminator be fooled

Title Don’t let your Discriminator be fooled
Authors Brady Zhou, Philipp Krähenbühl
Abstract Generative Adversarial Networks are one of the leading tools in generative modeling, image editing and content creation. However, they are hard to train as they require a delicate balancing act between two deep networks fighting a never ending duel. Some of the most promising adversarial models today minimize a Wasserstein objective. It is smoother and more stable to optimize. In this paper, we show that the Wasserstein distance is just one out of a large family of objective functions that yield these properties. By making the discriminator of a GAN robust to adversarial attacks we can turn any GAN objective into a smooth and stable loss. We experimentally show that any GAN objective, including Wasserstein GANs, benefit from adversarial robustness both quantitatively and qualitatively. The training additionally becomes more robust to suboptimal choices of hyperparameters, model architectures, or objective functions.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=HJE6X305Fm
PDF https://openreview.net/pdf?id=HJE6X305Fm
PWC https://paperswithcode.com/paper/dont-let-your-discriminator-be-fooled
Repo
Framework

Diversity and Depth in Per-Example Routing Models

Title Diversity and Depth in Per-Example Routing Models
Authors Prajit Ramachandran, Quoc V. Le
Abstract Routing models, a form of conditional computation where examples are routed through a subset of components in a larger network, have shown promising results in recent works. Surprisingly, routing models to date have lacked important properties, such as architectural diversity and large numbers of routing decisions. Both architectural diversity and routing depth can increase the representational power of a routing network. In this work, we address both of these deficiencies. We discuss the significance of architectural diversity in routing models, and explain the tradeoffs between capacity and optimization when increasing routing depth. In our experiments, we find that adding architectural diversity to routing models significantly improves performance, cutting the error rates of a strong baseline by 35% on an Omniglot setup. However, when scaling up routing depth, we find that modern routing techniques struggle with optimization. We conclude by discussing both the positive and negative results, and suggest directions for future research.
Tasks Multi-Task Learning, Omniglot
Published 2019-05-01
URL https://openreview.net/forum?id=BkxWJnC9tX
PDF https://openreview.net/pdf?id=BkxWJnC9tX
PWC https://paperswithcode.com/paper/diversity-and-depth-in-per-example-routing
Repo
Framework

Modeling Hierarchical Syntactic Structures in Morphological Processing

Title Modeling Hierarchical Syntactic Structures in Morphological Processing
Authors Yohei Oseki, Charles Yang, Alec Marantz
Abstract Sentences are represented as hierarchical syntactic structures, which have been successfully modeled in sentence processing. In contrast, despite the theoretical agreement on hierarchical syntactic structures within words, words have been argued to be computationally less complex than sentences and implemented by finite-state models as linear strings of morphemes, and even the psychological reality of morphemes has been denied. In this paper, extending the computational models employed in sentence processing to morphological processing, we performed a computational simulation experiment where, given incremental surprisal as a linking hypothesis, five computational models with different representational assumptions were evaluated against human reaction times in visual lexical decision experiments available from the English Lexicon Project (ELP), a {}shared task{''} in the morphological processing literature. The simulation experiment demonstrated that (i) {}amorphous{''} models without morpheme units underperformed relative to {}morphous{''} models, (ii) a computational model with hierarchical syntactic structures, Probabilistic Context-Free Grammar (PCFG), most accurately explained human reaction times, and (iii) this performance was achieved on top of surface frequency effects. These results strongly suggest that morphological processing tracks morphemes incrementally from left to right and parses them into hierarchical syntactic structures, contrary to {}amorphous{''} and finite-state models of morphological processing.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-2905/
PDF https://www.aclweb.org/anthology/W19-2905
PWC https://paperswithcode.com/paper/modeling-hierarchical-syntactic-structures-in
Repo
Framework

Weighted parsing for grammar-based language models

Title Weighted parsing for grammar-based language models
Authors Richard M{"o}rbitz, Heiko Vogler
Abstract We develop a general framework for weighted parsing which is built on top of grammar-based language models and employs flexible weight algebras. It generalizes previous work in that area (semiring parsing, weighted deductive parsing) and also covers applications outside the classical scope of parsing, e.g., algebraic dynamic programming. We show an algorithm which terminates and is correct for a large class of weighted grammar-based language models.
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-3108/
PDF https://www.aclweb.org/anthology/W19-3108
PWC https://paperswithcode.com/paper/weighted-parsing-for-grammar-based-language
Repo
Framework

Online Learning for Supervised Dimension Reduction

Title Online Learning for Supervised Dimension Reduction
Authors Ning Zhang, Qiang Wu
Abstract Online learning has attracted great attention due to the increasing demand for systems that have the ability of learning and evolving. When the data to be processed is also high dimensional and dimension reduction is necessary for visualization or prediction enhancement, online dimension reduction will play an essential role. The purpose of this paper is to propose new online learning approaches for supervised dimension reduction. Our first algorithm is motivated by adapting the sliced inverse regression (SIR), a pioneer and effective algorithm for supervised dimension reduction, and making it implementable in an incremental manner. The new algorithm, called incremental sliced inverse regression (ISIR), is able to update the subspace of significant factors with intrinsic lower dimensionality fast and efficiently when new observations come in. We also refine the algorithm by using an overlapping technique and develop an incremental overlapping sliced inverse regression (IOSIR) algorithm. We verify the effectiveness and efficiency of both algorithms by simulations and real data applications.
Tasks Dimensionality Reduction
Published 2019-05-01
URL https://openreview.net/forum?id=B1MUroRct7
PDF https://openreview.net/pdf?id=B1MUroRct7
PWC https://paperswithcode.com/paper/online-learning-for-supervised-dimension
Repo
Framework

Transfer Capsule Network for Aspect Level Sentiment Classification

Title Transfer Capsule Network for Aspect Level Sentiment Classification
Authors Zhuang Chen, Tieyun Qian
Abstract Aspect-level sentiment classification aims to determine the sentiment polarity of a sentence towards an aspect. Due to the high cost in annotation, the lack of aspect-level labeled data becomes a major obstacle in this area. On the other hand, document-level labeled data like reviews are easily accessible from online websites. These reviews encode sentiment knowledge in abundant contexts. In this paper, we propose a Transfer Capsule Network (TransCap) model for transferring document-level knowledge to aspect-level sentiment classification. To this end, we first develop an aspect routing approach to encapsulate the sentence-level semantic representations into semantic capsules from both the aspect-level and document-level data. We then extend the dynamic routing approach to adaptively couple the semantic capsules with the class capsules under the transfer learning framework. Experiments on SemEval datasets demonstrate the effectiveness of TransCap.
Tasks Sentiment Analysis, Transfer Learning
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1052/
PDF https://www.aclweb.org/anthology/P19-1052
PWC https://paperswithcode.com/paper/transfer-capsule-network-for-aspect-level
Repo
Framework

Accelerated Sparse Recovery Under Structured Measurements

Title Accelerated Sparse Recovery Under Structured Measurements
Authors Ke Li, Jitendra Malik
Abstract Extensive work on compressed sensing has yielded a rich collection of sparse recovery algorithms, each making different tradeoffs between recovery condition and computational efficiency. In this paper, we propose a unified framework for accelerating various existing sparse recovery algorithms without sacrificing recovery guarantees by exploiting structure in the measurement matrix. Unlike fast algorithms that are specific to particular choices of measurement matrices where the columns are Fourier or wavelet filters for example, the proposed approach works on a broad range of measurement matrices that satisfy a particular property. We precisely characterize this property, which quantifies how easy it is to accelerate sparse recovery for the measurement matrix in question. We also derive the time complexity of the accelerated algorithm, which is sublinear in the signal length in each iteration. Moreover, we present experimental results on real world data that demonstrate the effectiveness of the proposed approach in practice.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=ryxyHnR5tX
PDF https://openreview.net/pdf?id=ryxyHnR5tX
PWC https://paperswithcode.com/paper/accelerated-sparse-recovery-under-structured
Repo
Framework

Studying Laws of Semantic Divergence across Languages using Cognate Sets

Title Studying Laws of Semantic Divergence across Languages using Cognate Sets
Authors Ana Uban, Alina Maria Ciobanu, Liviu P. Dinu
Abstract Semantic divergence in related languages is a key concern of historical linguistics. Intra-lingual semantic shift has been previously studied in computational linguistics, but this can only provide a limited picture of the evolution of word meanings, which often develop in a multilingual environment. In this paper we investigate semantic change across languages by measuring the semantic distance of cognate words in multiple languages. By comparing current meanings of cognates in different languages, we hope to uncover information about their previous meanings, and about how they diverged in their respective languages from their common original etymon. We further study the properties of their semantic divergence, by analyzing how the features of words such as frequency and polysemy are related to the divergence in their meaning, and thus make the first steps towards formulating laws of cross-lingual semantic change.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4720/
PDF https://www.aclweb.org/anthology/W19-4720
PWC https://paperswithcode.com/paper/studying-laws-of-semantic-divergence-across
Repo
Framework

Multi-headed Architecture Based on BERT for Grammatical Errors Correction

Title Multi-headed Architecture Based on BERT for Grammatical Errors Correction
Authors Bohdan Didenko, Julia Shaptala
Abstract In this paper, we describe our approach to GEC using the BERT model for creation of encoded representation and some of our enhancements, namely, {``}Heads{''} are fully-connected networks which are used for finding the errors and later receive recommendation from the networks on dealing with a highlighted part of the sentence only. Among the main advantages of our solution is increasing the system productivity and lowering the time of processing while keeping the high accuracy of GEC results. |
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4426/
PDF https://www.aclweb.org/anthology/W19-4426
PWC https://paperswithcode.com/paper/multi-headed-architecture-based-on-bert-for
Repo
Framework

Proceedings of the Third Workshop on Natural Language Processing and Computational Social Science

Title Proceedings of the Third Workshop on Natural Language Processing and Computational Social Science
Authors
Abstract
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-2100/
PDF https://www.aclweb.org/anthology/W19-2100
PWC https://paperswithcode.com/paper/proceedings-of-the-third-workshop-on-natural
Repo
Framework

Towards Machine Reading for Interventions from Humanitarian-Assistance Program Literature

Title Towards Machine Reading for Interventions from Humanitarian-Assistance Program Literature
Authors Bonan Min, Yee Seng Chan, Haoling Qiu, Joshua Fasching
Abstract Solving long-lasting problems such as food insecurity requires a comprehensive understanding of interventions applied by governments and international humanitarian assistance organizations, and their results and consequences. Towards achieving this grand goal, a crucial first step is to extract past interventions and when and where they have been applied, from hundreds of thousands of reports automatically. In this paper, we developed a corpus annotated with interventions to foster research, and developed an information extraction system for extracting interventions and their location and time from text. We demonstrate early, very encouraging results on extracting interventions.
Tasks Reading Comprehension
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1680/
PDF https://www.aclweb.org/anthology/D19-1680
PWC https://paperswithcode.com/paper/towards-machine-reading-for-interventions
Repo
Framework
comments powered by Disqus