January 24, 2020

2262 words 11 mins read

Paper Group NANR 224

Paper Group NANR 224

Comparing Top-Down and Bottom-Up Neural Generative Dependency Models. Pardon the Interruption: Automatic Analysis of Gender and Competitive Turn-Taking in United States Supreme Court Hearings. Summarizing Legal Rulings: Comparative Experiments. A Recurrent BERT-based Model for Question Generation. Learning from Positive and Unlabeled Data with a Se …

Comparing Top-Down and Bottom-Up Neural Generative Dependency Models

Title Comparing Top-Down and Bottom-Up Neural Generative Dependency Models
Authors Austin Matthews, Graham Neubig, Chris Dyer
Abstract Recurrent neural network grammars generate sentences using phrase-structure syntax and perform very well on both parsing and language modeling. To explore whether generative dependency models are similarly effective, we propose two new generative models of dependency syntax. Both models use recurrent neural nets to avoid making explicit independence assumptions, but they differ in the order used to construct the trees: one builds the tree bottom-up and the other top-down, which profoundly changes the estimation problem faced by the learner. We evaluate the two models on three typologically different languages: English, Arabic, and Japanese. While both generative models improve parsing performance over a discriminative baseline, they are significantly less effective than non-syntactic LSTM language models. Surprisingly, little difference between the construction orders is observed for either parsing or language modeling.
Tasks Language Modelling
Published 2019-11-01
URL https://www.aclweb.org/anthology/K19-1022/
PDF https://www.aclweb.org/anthology/K19-1022
PWC https://paperswithcode.com/paper/comparing-top-down-and-bottom-up-neural
Repo
Framework

Pardon the Interruption: Automatic Analysis of Gender and Competitive Turn-Taking in United States Supreme Court Hearings

Title Pardon the Interruption: Automatic Analysis of Gender and Competitive Turn-Taking in United States Supreme Court Hearings
Authors Haley Lepp
Abstract The United States Supreme Court plays a key role in defining the legal basis for gender discrimination throughout the country, yet there are few checks on gender bias within the court itself. In conversational turn-taking, interruptions have been documented as a marker of bias between speakers of different genders. The goal of this study is to automatically differentiate between respectful and disrespectful conversational turns taken during official hearings, which could help in detecting bias and finding remediation techniques for discourse in the courtroom. In this paper, I present a corpus of turns annotated by legal professionals, and describe the design of a semi-supervised classifier that will use acoustic and lexical features to analyze turn-taking at scale. On completion of annotations, this classifier will be trained to extract the likelihood that turns are respectful or disrespectful for use in studies of speech trends.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/papers/W/W19/W19-3645/
PDF https://www.aclweb.org/anthology/W19-3645
PWC https://paperswithcode.com/paper/pardon-the-interruption-automatic-analysis-of
Repo
Framework
Title Summarizing Legal Rulings: Comparative Experiments
Authors Diego Feijo, Viviane Moreira
Abstract In the context of text summarization, texts in the legal domain have peculiarities related to their length and to their specialized vocabulary. Recent neural network-based approaches can achieve high-quality scores for text summarization. However, these approaches have been used mostly for generating very short abstracts for news articles. Thus, their applicability to the legal domain remains an open issue. In this work, we experimented with ten extractive and four abstractive models in a real dataset of legal rulings. These models were compared with an extractive baseline based on heuristics to select the most relevant parts of the text. Our results show that abstractive approaches significantly outperform extractive methods in terms of ROUGE scores.
Tasks Text Summarization
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-1036/
PDF https://www.aclweb.org/anthology/R19-1036
PWC https://paperswithcode.com/paper/summarizing-legal-rulings-comparative
Repo
Framework

A Recurrent BERT-based Model for Question Generation

Title A Recurrent BERT-based Model for Question Generation
Authors Ying-Hong Chan, Yao-Chung Fan
Abstract In this study, we investigate the employment of the pre-trained BERT language model to tackle question generation tasks. We introduce three neural architectures built on top of BERT for question generation tasks. The first one is a straightforward BERT employment, which reveals the defects of directly using BERT for text generation. Accordingly, we propose another two models by restructuring our BERT employment into a sequential manner for taking information from previous decoded results. Our models are trained and evaluated on the recent question-answering dataset SQuAD. Experiment results show that our best model yields state-of-the-art performance which advances the BLEU 4 score of the existing best models from 16.85 to 22.17.
Tasks Language Modelling, Question Answering, Question Generation, Text Generation
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5821/
PDF https://www.aclweb.org/anthology/D19-5821
PWC https://paperswithcode.com/paper/a-recurrent-bert-based-model-for-question
Repo
Framework

Learning from Positive and Unlabeled Data with a Selection Bias

Title Learning from Positive and Unlabeled Data with a Selection Bias
Authors Masahiro Kato, Takeshi Teshima, Junya Honda
Abstract We consider the problem of learning a binary classifier only from positive data and unlabeled data (PU learning). Recent methods of PU learning commonly assume that the labeled positive data are identically distributed as the unlabeled positive data. However, this assumption is unrealistic in many instances of PU learning because it fails to capture the existence of a selection bias in the labeling process. When the data has a selection bias, it is difficult to learn the Bayes optimal classifier by conventional methods of PU learning. In this paper, we propose a method to partially identify the classifier. The proposed algorithm learns a scoring function that preserves the order induced by the class posterior under mild assumptions, which can be used as a classifier by setting an appropriate threshold. Through experiments, we show that the method outperforms previous methods for PU learning on various real-world datasets.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=rJzLciCqKm
PDF https://openreview.net/pdf?id=rJzLciCqKm
PWC https://paperswithcode.com/paper/learning-from-positive-and-unlabeled-data-1
Repo
Framework

Diversity-aware Event Prediction based on a Conditional Variational Autoencoder with Reconstruction

Title Diversity-aware Event Prediction based on a Conditional Variational Autoencoder with Reconstruction
Authors Hirokazu Kiyomaru, Kazumasa Omura, Yugo Murawaki, Daisuke Kawahara, Sadao Kurohashi
Abstract Typical event sequences are an important class of commonsense knowledge. Formalizing the task as the generation of a next event conditioned on a current event, previous work in event prediction employs sequence-to-sequence (seq2seq) models. However, what can happen after a given event is usually diverse, a fact that can hardly be captured by deterministic models. In this paper, we propose to incorporate a conditional variational autoencoder (CVAE) into seq2seq for its ability to represent diverse next events as a probabilistic distribution. We further extend the CVAE-based seq2seq with a reconstruction mechanism to prevent the model from concentrating on highly typical events. To facilitate fair and systematic evaluation of the diversity-aware models, we also extend existing evaluation datasets by tying each current event to multiple next events. Experiments show that the CVAE-based models drastically outperform deterministic models in terms of precision and that the reconstruction mechanism improves the recall of CVAE-based models without sacrificing precision.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-6014/
PDF https://www.aclweb.org/anthology/D19-6014
PWC https://paperswithcode.com/paper/diversity-aware-event-prediction-based-on-a
Repo
Framework

Distributional Effects of Gender Contrasts Across Categories

Title Distributional Effects of Gender Contrasts Across Categories
Authors Timothee Mickus, Olivier Bonami, Denis Paperno
Abstract
Tasks
Published 2019-01-01
URL https://www.aclweb.org/anthology/W19-0118/
PDF https://www.aclweb.org/anthology/W19-0118
PWC https://paperswithcode.com/paper/distributional-effects-of-gender-contrasts
Repo
Framework

Case assignment in TSL syntax: a case study

Title Case assignment in TSL syntax: a case study
Authors Mai Ha Vu, Nazila Shafiei, Thomas Graf
Abstract
Tasks
Published 2019-01-01
URL https://www.aclweb.org/anthology/W19-0127/
PDF https://www.aclweb.org/anthology/W19-0127
PWC https://paperswithcode.com/paper/case-assignment-in-tsl-syntax-a-case-study
Repo
Framework

Are formal restrictions on crossing dependencies epiphenominal?

Title Are formal restrictions on crossing dependencies epiphenominal?
Authors Himanshu Yadav, Samar Husain, Richard Futrell
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-7802/
PDF https://www.aclweb.org/anthology/W19-7802
PWC https://paperswithcode.com/paper/are-formal-restrictions-on-crossing
Repo
Framework

Automatically Extracting Challenge Sets for Non-Local Phenomena in Neural Machine Translation

Title Automatically Extracting Challenge Sets for Non-Local Phenomena in Neural Machine Translation
Authors Leshem Choshen, Omri Abend
Abstract We show that the state-of-the-art Transformer MT model is not biased towards monotonic reordering (unlike previous recurrent neural network models), but that nevertheless, long-distance dependencies remain a challenge for the model. Since most dependencies are short-distance, common evaluation metrics will be little influenced by how well systems perform on them. We therefore propose an automatic approach for extracting challenge sets rich with long-distance dependencies, and argue that evaluation using this methodology provides a complementary perspective on system performance. To support our claim, we compile challenge sets for English-German and German-English, which are much larger than any previously released challenge set for MT. The extracted sets are large enough to allow reliable automatic evaluation, which makes the proposed approach a scalable and practical solution for evaluating MT performance on the long-tail of syntactic phenomena.
Tasks Machine Translation
Published 2019-11-01
URL https://www.aclweb.org/anthology/K19-1028/
PDF https://www.aclweb.org/anthology/K19-1028
PWC https://paperswithcode.com/paper/automatically-extracting-challenge-sets-for-1
Repo
Framework

Distributional Semantics Meets Construction Grammar. towards a Unified Usage-Based Model of Grammar and Meaning

Title Distributional Semantics Meets Construction Grammar. towards a Unified Usage-Based Model of Grammar and Meaning
Authors Giulia Rambelli, Emmanuele Chersoni, Philippe Blache, Chu-Ren Huang, Aless Lenci, ro
Abstract In this paper, we propose a new type of semantic representation of Construction Grammar that combines constructions with the vector representations used in Distributional Semantics. We introduce a new framework, Distributional Construction Grammar, where grammar and meaning are systematically modeled from language use, and finally, we discuss the kind of contributions that distributional models can provide to CxG representation from a linguistic and cognitive perspective.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3312/
PDF https://www.aclweb.org/anthology/W19-3312
PWC https://paperswithcode.com/paper/distributional-semantics-meets-construction
Repo
Framework

Detecting Subevents using Discourse and Narrative Features

Title Detecting Subevents using Discourse and Narrative Features
Authors Mohammed Aldawsari, Mark Finlayson
Abstract Recognizing the internal structure of events is a challenging language processing task of great importance for text understanding. We present a supervised model for automatically identifying when one event is a subevent of another. Building on prior work, we introduce several novel features, in particular discourse and narrative features, that significantly improve upon prior state-of-the-art performance. Error analysis further demonstrates the utility of these features. We evaluate our model on the only two annotated corpora with event hierarchies: HiEve and the Intelligence Community corpus. No prior system has been evaluated on both corpora. Our model outperforms previous systems on both corpora, achieving 0.74 BLANC F1 on the Intelligence Community corpus and 0.70 F1 on the HiEve corpus, respectively a 15 and 5 percentage point improvement over previous models.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1471/
PDF https://www.aclweb.org/anthology/P19-1471
PWC https://paperswithcode.com/paper/detecting-subevents-using-discourse-and
Repo
Framework

Ensemble Deep Manifold Similarity Learning Using Hard Proxies

Title Ensemble Deep Manifold Similarity Learning Using Hard Proxies
Authors Nicolas Aziere, Sinisa Todorovic
Abstract This paper is about learning deep representations of images such that images belonging to the same class have more similar representations than those belonging to different classes. For this goal, prior work typically uses the triplet or N-pair loss, specified in terms of either l2-distances or dot-products between deep features. However, such formulations seem poorly suited to the highly non-Euclidean deep feature space. Our first contribution is in specifying the N-pair loss in terms of manifold similarities between deep features. We introduce a new time- and memory-efficient method for estimating the manifold similarities by using a closed-form convergence solution of the Random Walk algorithm. Our efficiency comes, in part, from following the recent work that randomly partitions the deep feature space, and expresses image distances via representatives of the resulting subspaces, a.k.a. proxies. Our second contribution is aimed at reducing overfitting by estimating hard proxies that are as close to one another as possible, but remain in their respective subspaces. Our evaluation demonstrates that we outperform the state of the art in both image retrieval and clustering on the benchmark CUB-200-2011, Cars196, and Stanford Online Products datasets.
Tasks Image Retrieval
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Aziere_Ensemble_Deep_Manifold_Similarity_Learning_Using_Hard_Proxies_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Aziere_Ensemble_Deep_Manifold_Similarity_Learning_Using_Hard_Proxies_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/ensemble-deep-manifold-similarity-learning
Repo
Framework

Quantifiers in a Multimodal World: Hallucinating Vision with Language and Sound

Title Quantifiers in a Multimodal World: Hallucinating Vision with Language and Sound
Authors Alberto Testoni, S Pezzelle, ro, Raffaella Bernardi
Abstract Inspired by the literature on multisensory integration, we develop a computational model to ground quantifiers in perception. The model learns to pick, out of nine quantifiers ({}few{'}, {}many{'}, {}all{'}, etc.), the one that is more likely to describe the percent of animals in a visual-auditory input containing both animals and artifacts. We show that relying on concurrent sensory inputs increases model performance on the quantification task. Moreover, we evaluate the model in a situation in which only the auditory modality is given, while the visual one is {}hallucinanted{'} either from the auditory input itself or from a linguistic caption describing the quantity of entities in the auditory input. This way, the model exploits prior associations between modalities. We show that the model profits from the prior knowledge and outperforms the auditory-only setting.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-2912/
PDF https://www.aclweb.org/anthology/W19-2912
PWC https://paperswithcode.com/paper/quantifiers-in-a-multimodal-world
Repo
Framework

YouMakeup: A Large-Scale Domain-Specific Multimodal Dataset for Fine-Grained Semantic Comprehension

Title YouMakeup: A Large-Scale Domain-Specific Multimodal Dataset for Fine-Grained Semantic Comprehension
Authors Weiying Wang, Yongcheng Wang, Shizhe Chen, Qin Jin
Abstract Multimodal semantic comprehension has attracted increasing research interests recently such as visual question answering and caption generation. However, due to the data limitation, fine-grained semantic comprehension has not been well investigated, which requires to capture semantic details of multimodal contents. In this work, we introduce {``}YouMakeup{''}, a large-scale multimodal instructional video dataset to support fine-grained semantic comprehension research in specific domain. YouMakeup contains 2,800 videos from YouTube, spanning more than 420 hours in total. Each video is annotated with a sequence of natural language descriptions for instructional steps, grounded in temporal video range and spatial facial areas. The annotated steps in a video involve subtle difference in actions, products and regions, which requires fine-grained understanding and reasoning both temporally and spatially. In order to evaluate models{'} ability for fined-grained comprehension, we further propose two groups of tasks including generation tasks and visual question answering from different aspects. We also establish a baseline of step caption generation for future comparison. The dataset will be publicly available at https://github. com/AIM3-RUC/YouMakeup to support research investigation in fine-grained semantic comprehension. |
Tasks Question Answering, Visual Question Answering
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1517/
PDF https://www.aclweb.org/anthology/D19-1517
PWC https://paperswithcode.com/paper/youmakeup-a-large-scale-domain-specific
Repo
Framework
comments powered by Disqus