January 25, 2020

2393 words 12 mins read

Paper Group NANR 20

Paper Group NANR 20

PyOpenDial: A Python-based Domain-Independent Toolkit for Developing Spoken Dialogue Systems with Probabilistic Rules. Capturing Dialogue State Variable Dependencies with an Energy-based Neural Dialogue State Tracker. SyntaxFest 2019 Invited talk - Dependency distance minimization: facts, theory and predictions. Capsule Network with Interactive Att …

PyOpenDial: A Python-based Domain-Independent Toolkit for Developing Spoken Dialogue Systems with Probabilistic Rules

Title PyOpenDial: A Python-based Domain-Independent Toolkit for Developing Spoken Dialogue Systems with Probabilistic Rules
Authors Youngsoo Jang, Jongmin Lee, Jaeyoung Park, Kyeng-Hun Lee, Pierre Lison, Kee-Eung Kim
Abstract We present PyOpenDial, a Python-based domain-independent, open-source toolkit for spoken dialogue systems. Recent advances in core components of dialogue systems, such as speech recognition, language understanding, dialogue management, and language generation, harness deep learning to achieve state-of-the-art performance. The original OpenDial, implemented in Java, provides a plugin architecture to integrate external modules, but lacks Python bindings, making it difficult to interface with popular deep learning frameworks such as Tensorflow or PyTorch. To this end, we re-implemented OpenDial in Python and extended the toolkit with a number of novel functionalities for neural dialogue state tracking and action planning. We describe the overall architecture and its extensions, and illustrate their use on an example where the system response model is implemented with a recurrent neural network.
Tasks Dialogue Management, Dialogue State Tracking, Speech Recognition, Spoken Dialogue Systems, Text Generation
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-3032/
PDF https://www.aclweb.org/anthology/D19-3032
PWC https://paperswithcode.com/paper/pyopendial-a-python-based-domain-independent
Repo
Framework

Capturing Dialogue State Variable Dependencies with an Energy-based Neural Dialogue State Tracker

Title Capturing Dialogue State Variable Dependencies with an Energy-based Neural Dialogue State Tracker
Authors Anh Duong Trinh, Robert J. Ross, John D. Kelleher
Abstract Dialogue state tracking requires the population and maintenance of a multi-slot frame representation of the dialogue state. Frequently, dialogue state tracking systems assume independence between slot values within a frame. In this paper we argue that treating the prediction of each slot value as an independent prediction task may ignore important associations between the slot values, and, consequently, we argue that treating dialogue state tracking as a structured prediction problem can help to improve dialogue state tracking performance. To support this argument, the research presented in this paper is structured into three stages: (i) analyzing variable dependencies in dialogue data; (ii) applying an energy-based methodology to model dialogue state tracking as a structured prediction task; and (iii) evaluating the impact of inter-slot relationships on model performance. Overall we demonstrate that modelling the associations between target slots with an energy-based formalism improves dialogue state tracking performance in a number of ways.
Tasks Dialogue State Tracking, Structured Prediction
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-5910/
PDF https://www.aclweb.org/anthology/W19-5910
PWC https://paperswithcode.com/paper/capturing-dialogue-state-variable
Repo
Framework

SyntaxFest 2019 Invited talk - Dependency distance minimization: facts, theory and predictions

Title SyntaxFest 2019 Invited talk - Dependency distance minimization: facts, theory and predictions
Authors Ramon Ferrer-i-Cancho
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-7901/
PDF https://www.aclweb.org/anthology/W19-7901
PWC https://paperswithcode.com/paper/syntaxfest-2019-invited-talk-dependency
Repo
Framework

Capsule Network with Interactive Attention for Aspect-Level Sentiment Classification

Title Capsule Network with Interactive Attention for Aspect-Level Sentiment Classification
Authors Chunning Du, Haifeng Sun, Jingyu Wang, Qi Qi, Jianxin Liao, Tong Xu, Ming Liu
Abstract Aspect-level sentiment classification is a crucial task for sentiment analysis, which aims to identify the sentiment polarities of specific targets in their context. The main challenge comes from multi-aspect sentences, which express multiple sentiment polarities towards different targets, resulting in overlapped feature representation. However, most existing neural models tend to utilize static pooling operation or attention mechanism to identify sentimental words, which therefore insufficient for dealing with overlapped features. To solve this problem, we propose to utilize capsule network to construct vector-based feature representation and cluster features by an EM routing algorithm. Furthermore, interactive attention mechanism is introduced in the capsule routing procedure to model the semantic relationship between aspect terms and context. The iterative routing also enables encoding sentence from a global perspective. Experimental results on three datasets show that our proposed model achieves state-of-the-art performance.
Tasks Sentiment Analysis
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1551/
PDF https://www.aclweb.org/anthology/D19-1551
PWC https://paperswithcode.com/paper/capsule-network-with-interactive-attention
Repo
Framework

CBNU System for SIGMORPHON 2019 Shared Task 2: a Pipeline Model

Title CBNU System for SIGMORPHON 2019 Shared Task 2: a Pipeline Model
Authors Uygun Shadikhodjaev, Jae Sung Lee
Abstract In this paper we describe our system for morphological analysis and lemmatization in context, using a transformer-based sequence to sequence model and a biaffine attention based BiLSTM model. First, a lemma is produced for a given word, and then both the lemma and the given word are used for morphological analysis. We also make use of character level word encodings and trainable encodings to improve accuracy. Overall, our system ranked fifth in lemmatization and sixth in morphological accuracy among twelve systems, and demonstrated considerable improvements over the baseline in morphological analysis.
Tasks Lemmatization, Morphological Analysis
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4204/
PDF https://www.aclweb.org/anthology/W19-4204
PWC https://paperswithcode.com/paper/cbnu-system-for-sigmorphon-2019-shared-task-2
Repo
Framework

Leveraging Adjective-Noun Phrasing Knowledge for Comparison Relation Prediction in Text-to-SQL

Title Leveraging Adjective-Noun Phrasing Knowledge for Comparison Relation Prediction in Text-to-SQL
Authors Haoyan Liu, Lei Fang, Qian Liu, Bei Chen, Jian-Guang Lou, Zhoujun Li
Abstract One key component in text-to-SQL is to predict the comparison relations between columns and their values. To the best of our knowledge, no existing models explicitly introduce external common knowledge to address this problem, thus their capabilities of predicting comparison relations are limited beyond training data. In this paper, we propose to leverage adjective-noun phrasing knowledge mined from the web to predict the comparison relations in text-to-SQL. Experimental results on both the original and the re-split Spider dataset show that our approach achieves significant improvement over state-of-the-art methods on comparison relation prediction.
Tasks Text-To-Sql
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1356/
PDF https://www.aclweb.org/anthology/D19-1356
PWC https://paperswithcode.com/paper/leveraging-adjective-noun-phrasing-knowledge
Repo
Framework

Occlusion Robust Face Recognition Based on Mask Learning With Pairwise Differential Siamese Network

Title Occlusion Robust Face Recognition Based on Mask Learning With Pairwise Differential Siamese Network
Authors Lingxue Song, Dihong Gong, Zhifeng Li, Changsong Liu, Wei Liu
Abstract Deep Convolutional Neural Networks (CNNs) have been pushing the frontier of face recognition over past years. However, existing CNN models are far less accurate when handling partially occluded faces. These general face models generalize poorly for occlusions on variable facial areas. Inspired by the fact that human visual system explicitly ignores the occlusion and only focuses on the non-occluded facial areas, we propose a mask learning strategy to find and discard corrupted feature elements from recognition. A mask dictionary is firstly established by exploiting the differences between the top conv features of occluded and occlusion-free face pairs using innovatively designed pairwise differential siamese network (PDSN). Each item of this dictionary captures the correspondence between occluded facial areas and corrupted feature elements, which is named Feature Discarding Mask (FDM). When dealing with a face image with random partial occlusions, we generate its FDM by combining relevant dictionary items and then multiply it with the original features to eliminate those corrupted feature elements from recognition. Comprehensive experiments on both synthesized and realistic occluded face datasets show that the proposed algorithm significantly outperforms the state-of-the-art systems.
Tasks Face Recognition, Robust Face Recognition
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Song_Occlusion_Robust_Face_Recognition_Based_on_Mask_Learning_With_Pairwise_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Song_Occlusion_Robust_Face_Recognition_Based_on_Mask_Learning_With_Pairwise_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/occlusion-robust-face-recognition-based-on-1
Repo
Framework

Distraction-Aware Shadow Detection

Title Distraction-Aware Shadow Detection
Authors Quanlong Zheng, Xiaotian Qiao, Ying Cao, Rynson W.H. Lau
Abstract Shadow detection is an important and challenging task for scene understanding. Despite promising results from recent deep learning based methods. Existing works still struggle with ambiguous cases where the visual appearances of shadow and non-shadow regions are similar (referred to as distraction in our context). In this paper, we propose a Distraction-aware Shadow Detection Network (DSDNet) by explicitly learning and integrating the semantics of visual distraction regions in an end-to-end framework. At the core of our framework is a novel standalone, differentiable Distraction-aware Shadow (DS) module, which allows us to learn distraction-aware, discriminative features for robust shadow detection, by explicitly predicting false positives and false negatives. We conduct extensive experiments on three public shadow detection datasets, SBU, UCF and ISTD, to evaluate our method. Experimental results demonstrate that our model can boost shadow detection performance, by effectively suppressing the detection of false positives and false negatives, achieving state-of-the-art results.
Tasks Scene Understanding, Shadow Detection
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Zheng_Distraction-Aware_Shadow_Detection_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Zheng_Distraction-Aware_Shadow_Detection_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/distraction-aware-shadow-detection
Repo
Framework

Corpus Building for Low Resource Languages in the DARPA LORELEI Program

Title Corpus Building for Low Resource Languages in the DARPA LORELEI Program
Authors Jennifer Tracey, Stephanie Strassel, Ann Bies, Zhiyi Song, Michael Arrigo, Kira Griffitt, Dana Delgado, Dave Graff, Seth Kulick, Justin Mott, Neil Kuster
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-6808/
PDF https://www.aclweb.org/anthology/W19-6808
PWC https://paperswithcode.com/paper/corpus-building-for-low-resource-languages-in
Repo
Framework

Transductive Learning of Neural Language Models for Syntactic and Semantic Analysis

Title Transductive Learning of Neural Language Models for Syntactic and Semantic Analysis
Authors Hiroki Ouchi, Jun Suzuki, Kentaro Inui
Abstract In transductive learning, an unlabeled test set is used for model training. Although this setting deviates from the common assumption of a completely unseen test set, it is applicable in many real-world scenarios, wherein the texts to be processed are known in advance. However, despite its practical advantages, transductive learning is underexplored in natural language processing. Here we conduct an empirical study of transductive learning for neural models and demonstrate its utility in syntactic and semantic tasks. Specifically, we fine-tune language models (LMs) on an unlabeled test set to obtain test-set-specific word representations. Through extensive experiments, we demonstrate that despite its simplicity, transductive LM fine-tuning consistently improves state-of-the-art neural models in in-domain and out-of-domain settings.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1379/
PDF https://www.aclweb.org/anthology/D19-1379
PWC https://paperswithcode.com/paper/transductive-learning-of-neural-language
Repo
Framework

Ranking Generated Summaries by Correctness: An Interesting but Challenging Application for Natural Language Inference

Title Ranking Generated Summaries by Correctness: An Interesting but Challenging Application for Natural Language Inference
Authors Tobias Falke, Leonardo F. R. Ribeiro, Prasetya Ajie Utama, Ido Dagan, Iryna Gurevych
Abstract While recent progress on abstractive summarization has led to remarkably fluent summaries, factual errors in generated summaries still severely limit their use in practice. In this paper, we evaluate summaries produced by state-of-the-art models via crowdsourcing and show that such errors occur frequently, in particular with more abstractive models. We study whether textual entailment predictions can be used to detect such errors and if they can be reduced by reranking alternative predicted summaries. That leads to an interesting downstream application for entailment models. In our experiments, we find that out-of-the-box entailment models trained on NLI datasets do not yet offer the desired performance for the downstream task and we therefore release our annotations as additional test data for future extrinsic evaluations of NLI.
Tasks Abstractive Text Summarization, Natural Language Inference
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1213/
PDF https://www.aclweb.org/anthology/P19-1213
PWC https://paperswithcode.com/paper/ranking-generated-summaries-by-correctness-an
Repo
Framework

Domain Adaptation with BERT-based Domain Classification and Data Selection

Title Domain Adaptation with BERT-based Domain Classification and Data Selection
Authors Xiaofei Ma, Peng Xu, Zhiguo Wang, Ramesh Nallapati, Bing Xiang
Abstract The performance of deep neural models can deteriorate substantially when there is a domain shift between training and test data. For example, the pre-trained BERT model can be easily fine-tuned with just one additional output layer to create a state-of-the-art model for a wide range of tasks. However, the fine-tuned BERT model suffers considerably at zero-shot when applied to a different domain. In this paper, we present a novel two-step domain adaptation framework based on curriculum learning and domain-discriminative data selection. The domain adaptation is conducted in a mostly unsupervised manner using a small target domain validation set for hyper-parameter tuning. We tested the framework on four large public datasets with different domain similarities and task types. Our framework outperforms a popular discrepancy-based domain adaptation method on most transfer tasks while consuming only a fraction of the training budget.
Tasks Domain Adaptation
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-6109/
PDF https://www.aclweb.org/anthology/D19-6109
PWC https://paperswithcode.com/paper/domain-adaptation-with-bert-based-domain
Repo
Framework

SpatialNet: A Declarative Resource for Spatial Relations

Title SpatialNet: A Declarative Resource for Spatial Relations
Authors Morgan Ulinski, Bob Coyne, Julia Hirschberg
Abstract This paper introduces SpatialNet, a novel resource which links linguistic expressions to actual spatial configurations. SpatialNet is based on FrameNet (Ruppenhofer et al., 2016) and VigNet (Coyne et al., 2011), two resources which use frame semantics to encode lexical meaning. SpatialNet uses a deep semantic representation of spatial relations to provide a formal description of how a language expresses spatial information. This formal representation of the lexical semantics of spatial language also provides a consistent way to represent spatial meaning across multiple languages. In this paper, we describe the structure of SpatialNet, with examples from English and German. We also show how SpatialNet can be combined with other existing NLP tools to create a text-to-scene system for a language.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-1607/
PDF https://www.aclweb.org/anthology/W19-1607
PWC https://paperswithcode.com/paper/spatialnet-a-declarative-resource-for-spatial
Repo
Framework

Tagging modality in Oceanic languages of Melanesia

Title Tagging modality in Oceanic languages of Melanesia
Authors Annika Tjuka, Lena Wei{\ss}mann, Kilu von Prince
Abstract Primary data from small, low-resource languages of Oceania have only recently become available through language documentation. In our study, we explore corpus data of five Oceanic languages of Melanesia which are known to be mood-prominent (in the sense of Bhat, 1999). In order to find out more about tense, aspect, modality, and polarity, we tagged these categories in a subset of our corpora. For the category of modality, we developed a novel tag set (MelaTAMP, 2017), which categorizes clauses into factual, possible, and counterfactual. Based on an analysis of the inter-annotator consistency, we argue that our tag set for the modal domain is efficient for our subject languages and might be useful for other languages and purposes.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4008/
PDF https://www.aclweb.org/anthology/W19-4008
PWC https://paperswithcode.com/paper/tagging-modality-in-oceanic-languages-of
Repo
Framework

A preconditioned accelerated stochastic gradient descent algorithm

Title A preconditioned accelerated stochastic gradient descent algorithm
Authors Alexandru Onose, Seyed Iman Mossavat, Henk-Jan H. Smilde
Abstract We propose a preconditioned accelerated stochastic gradient method suitable for large scale optimization. We derive sufficient convergence conditions for the minimization of convex functions using a generic class of diagonal preconditioners and provide a formal convergence proof based on a framework originally used for on-line learning. Inspired by recent popular adaptive per-feature algorithms, we propose a specific preconditioner based on the second moment of the gradient. The sufficient convergence conditions motivate a critical adaptation of the per-feature updates in order to ensure convergence. We show empirical results for the minimization of convex and non-convex cost functions, in the context of neural network training. The method compares favorably with respect to current, first order, stochastic optimization methods.
Tasks Stochastic Optimization
Published 2019-05-01
URL https://openreview.net/forum?id=HJgODj05KX
PDF https://openreview.net/pdf?id=HJgODj05KX
PWC https://paperswithcode.com/paper/a-preconditioned-accelerated-stochastic
Repo
Framework
comments powered by Disqus