October 15, 2019

2471 words 12 mins read

Paper Group NANR 67

Paper Group NANR 67

Grounding Semantic Roles in Images. DNN Multimodal Fusion Techniques for Predicting Video Sentiment. Deconvolutional Time Series Regression: A Technique for Modeling Temporally Diffuse Effects. Is this Sentence Difficult? Do you Agree?. Seeing Temporal Modulation of Lights From Standard Cameras. Neural Transition Based Parsing of Web Queries: An En …

Grounding Semantic Roles in Images

Title Grounding Semantic Roles in Images
Authors Carina Silberer, Manfred Pinkal
Abstract We address the task of visual semantic role labeling (vSRL), the identification of the participants of a situation or event in a visual scene, and their labeling with their semantic relations to the event or situation. We render candidate participants as image regions of objects, and train a model which learns to ground roles in the regions which depict the corresponding participant. Experimental results demonstrate that we can train a vSRL model without reliance on prohibitive image-based role annotations, by utilizing noisy data which we extract automatically from image captions using a linguistic SRL system. Furthermore, our model induces frame{—}semantic visual representations, and their comparison to previous work on supervised visual verb sense disambiguation yields overall better results.
Tasks Image Captioning, Question Answering, Semantic Role Labeling
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1282/
PDF https://www.aclweb.org/anthology/D18-1282
PWC https://paperswithcode.com/paper/grounding-semantic-roles-in-images
Repo
Framework

DNN Multimodal Fusion Techniques for Predicting Video Sentiment

Title DNN Multimodal Fusion Techniques for Predicting Video Sentiment
Authors Jennifer Williams, Ramona Comanescu, Oana Radu, Leimin Tian
Abstract We present our work on sentiment prediction using the benchmark MOSI dataset from the CMU-MultimodalDataSDK. Previous work on multimodal sentiment analysis have been focused on input-level feature fusion or decision-level fusion for multimodal fusion. Here, we propose an intermediate-level feature fusion, which merges weights from each modality (audio, video, and text) during training with subsequent additional training. Moreover, we tested principle component analysis (PCA) for feature selection. We found that applying PCA increases unimodal performance, and multimodal fusion outperforms unimodal models. Our experiments show that our proposed intermediate-level feature fusion outperforms other fusion techniques, and it achieves the best performance with an overall binary accuracy of 74.0{%} on video+text modalities. Our work also improves feature selection for unimodal sentiment analysis, while proposing a novel and effective multimodal fusion architecture for this task.
Tasks Feature Selection, Multimodal Sentiment Analysis, Sentiment Analysis
Published 2018-07-01
URL https://www.aclweb.org/anthology/W18-3309/
PDF https://www.aclweb.org/anthology/W18-3309
PWC https://paperswithcode.com/paper/dnn-multimodal-fusion-techniques-for
Repo
Framework

Deconvolutional Time Series Regression: A Technique for Modeling Temporally Diffuse Effects

Title Deconvolutional Time Series Regression: A Technique for Modeling Temporally Diffuse Effects
Authors Cory Shain, William Schuler
Abstract Researchers in computational psycholinguistics frequently use linear models to study time series data generated by human subjects. However, time series may violate the assumptions of these models through temporal diffusion, where stimulus presentation has a lingering influence on the response as the rest of the experiment unfolds. This paper proposes a new statistical model that borrows from digital signal processing by recasting the predictors and response as convolutionally-related signals, using recent advances in machine learning to fit latent impulse response functions (IRFs) of arbitrary shape. A synthetic experiment shows successful recovery of true latent IRFs, and psycholinguistic experiments reveal plausible, replicable, and fine-grained estimates of latent temporal dynamics, with comparable or improved prediction quality to widely-used alternatives.
Tasks Time Series
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1288/
PDF https://www.aclweb.org/anthology/D18-1288
PWC https://paperswithcode.com/paper/deconvolutional-time-series-regression-a
Repo
Framework

Is this Sentence Difficult? Do you Agree?

Title Is this Sentence Difficult? Do you Agree?
Authors Dominique Brunato, Lorenzo De Mattei, Felice Dell{'}Orletta, Benedetta Iavarone, Giulia Venturi
Abstract In this paper, we present a crowdsourcing-based approach to model the human perception of sentence complexity. We collect a large corpus of sentences rated with judgments of complexity for two typologically-different languages, Italian and English. We test our approach in two experimental scenarios aimed to investigate the contribution of a wide set of lexical, morpho-syntactic and syntactic phenomena in predicting i) the degree of agreement among annotators independently from the assigned judgment and ii) the perception of sentence complexity.
Tasks Language Acquisition
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1289/
PDF https://www.aclweb.org/anthology/D18-1289
PWC https://paperswithcode.com/paper/is-this-sentence-difficult-do-you-agree
Repo
Framework

Seeing Temporal Modulation of Lights From Standard Cameras

Title Seeing Temporal Modulation of Lights From Standard Cameras
Authors Naoki Sakakibara, Fumihiko Sakaue, Jun Sato
Abstract In this paper, we propose a novel method for measuring the temporal modulation of lights by using off-the-shelf cameras. In particular, we show that the invisible flicker patterns of various lights such as fluorescent lights can be measured by a simple combination of an off-the-shelf camera and any moving object with specular reflection. Unlike the existing methods, we do not need high speed cameras nor specially designed coded exposure cameras. Based on the extracted flicker patterns of environment lights, we also propose an efficient method for deblurring motion blurs in images. The proposed method enables us to deblur images with better frequency characteristics, which are induced by the flicker patterns of environment lights. The real image experiments show the efficiency of the proposed method.
Tasks Deblurring
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/Sakakibara_Seeing_Temporal_Modulation_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/Sakakibara_Seeing_Temporal_Modulation_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/seeing-temporal-modulation-of-lights-from
Repo
Framework

Neural Transition Based Parsing of Web Queries: An Entity Based Approach

Title Neural Transition Based Parsing of Web Queries: An Entity Based Approach
Authors Rivka Malca, Roi Reichart
Abstract Web queries with question intent manifest a complex syntactic structure and the processing of this structure is important for their interpretation. Pinter et al. (2016) has formalized the grammar of these queries and proposed semi-supervised algorithms for the adaptation of parsers originally designed to parse according to the standard dependency grammar, so that they can account for the unique forest grammar of queries. However, their algorithms rely on resources typically not available outside of big web corporates. We propose a new BiLSTM query parser that: (1) Explicitly accounts for the unique grammar of web queries; and (2) Utilizes named entity (NE) information from a BiLSTM NE tagger, that can be jointly trained with the parser. In order to train our model we annotate the query treebank of Pinter et al. (2016) with NEs. When trained on 2500 annotated queries our parser achieves UAS of 83.5{%} and segmentation F1-score of 84.5, substantially outperforming existing state-of-the-art parsers.
Tasks Community Question Answering, Question Answering
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1290/
PDF https://www.aclweb.org/anthology/D18-1290
PWC https://paperswithcode.com/paper/neural-transition-based-parsing-of-web
Repo
Framework

Incremental Computation of Infix Probabilities for Probabilistic Finite Automata

Title Incremental Computation of Infix Probabilities for Probabilistic Finite Automata
Authors Marco Cognetta, Yo-Sub Han, Soon Chan Kwon
Abstract In natural language processing, a common task is to compute the probability of a phrase appearing in a document or to calculate the probability of all phrases matching a given pattern. For instance, one computes affix (prefix, suffix, infix, etc.) probabilities of a string or a set of strings with respect to a probability distribution of patterns. The problem of computing infix probabilities of strings when the pattern distribution is given by a probabilistic context-free grammar or by a probabilistic finite automaton is already solved, yet it was open to compute the infix probabilities in an incremental manner. The incremental computation is crucial when a new query is built from a previous query. We tackle this problem and suggest a method that computes infix probabilities incrementally for probabilistic finite automata by representing all the probabilities of matching strings as a series of transition matrix calculations. We show that the proposed approach is theoretically faster than the previous method and, using real world data, demonstrate that our approach has vastly better performance in practice.
Tasks
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1293/
PDF https://www.aclweb.org/anthology/D18-1293
PWC https://paperswithcode.com/paper/incremental-computation-of-infix
Repo
Framework

Last Words: What Can Be Accomplished with the State of the Art in Information Extraction? A Personal View

Title Last Words: What Can Be Accomplished with the State of the Art in Information Extraction? A Personal View
Authors Ralph Weischedel, Elizabeth Boschee
Abstract Though information extraction (IE) research has more than a 25-year history, F1 scores remain low. Thus, one could question continued investment in IE research. In this article, we present three applications where information extraction of entities, relations, and/or events has been used, and note the common features that seem to have led to success. We also identify key research challenges whose solution seems essential for broader successes. Because a few practical deployments already exist and because breakthroughs on particular challenges would greatly broaden the technology{'}s deployment, further R and D investments are justified.
Tasks Named Entity Recognition
Published 2018-12-01
URL https://www.aclweb.org/anthology/J18-4004/
PDF https://www.aclweb.org/anthology/J18-4004
PWC https://paperswithcode.com/paper/last-words-what-can-be-accomplished-with-the
Repo
Framework

Implicit Subjective and Sentimental Usages in Multi-sense Word Embeddings

Title Implicit Subjective and Sentimental Usages in Multi-sense Word Embeddings
Authors Yuqi Sun, Haoyue Shi, Junfeng Hu
Abstract In multi-sense word embeddings, contextual variations in corpus may cause a univocal word to be embedded into different sense vectors. Shi et al. (2016) show that this kind of \textit{pseudo multi-senses} can be eliminated by linear transformations. In this paper, we show that \textit{pseudo multi-senses} may come from a uniform and meaningful phenomenon such as subjective and sentimental usage, though they are seemingly redundant. In this paper, we present an unsupervised algorithm to find a linear transformation which can minimize the transformed distance of a group of sense pairs. The major shrinking direction of this transformation is found to be related with subjective shift. Therefore, we can not only eliminate \textit{pseudo multi-senses} in multisense embeddings, but also identify these subjective senses and tag the subjective and sentimental usage of words in the corpus automatically.
Tasks Word Embeddings
Published 2018-10-01
URL https://www.aclweb.org/anthology/W18-6203/
PDF https://www.aclweb.org/anthology/W18-6203
PWC https://paperswithcode.com/paper/implicit-subjective-and-sentimental-usages-in
Repo
Framework

Sentiment-Stance-Specificity (SSS) Dataset: Identifying Support-based Entailment among Opinions.

Title Sentiment-Stance-Specificity (SSS) Dataset: Identifying Support-based Entailment among Opinions.
Authors Pavithra Rajendran, Danushka Bollegala, Simon Parsons
Abstract
Tasks Argument Mining, Natural Language Inference
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1099/
PDF https://www.aclweb.org/anthology/L18-1099
PWC https://paperswithcode.com/paper/sentiment-stance-specificity-sss-dataset
Repo
Framework

A mostly unlexicalized model for recognizing textual entailment

Title A mostly unlexicalized model for recognizing textual entailment
Authors Mithun Paul, Rebecca Sharp, Mihai Surdeanu
Abstract Many approaches to automatically recognizing entailment relations have employed classifiers over hand engineered lexicalized features, or deep learning models that implicitly capture lexicalization through word embeddings. This reliance on lexicalization may complicate the adaptation of these tools between domains. For example, such a system trained in the news domain may learn that a sentence like {``}Palestinians recognize Texas as part of Mexico{''} tends to be unsupported, but this fact (and its corresponding lexicalized cues) have no value in, say, a scientific domain. To mitigate this dependence on lexicalized information, in this paper we propose a model that reads two sentences, from any given domain, to determine entailment without using lexicalized features. Instead our model relies on features that are either unlexicalized or are domain independent such as proportion of negated verbs, antonyms, or noun overlap. In its current implementation, this model does not perform well on the FEVER dataset, due to two reasons. First, for the information retrieval portion of the task we used the baseline system provided, since this was not the aim of our project. Second, this is work in progress and we still are in the process of identifying more features and gradually increasing the accuracy of our model. In the end, we hope to build a generic end-to-end classifier, which can be used in a domain outside the one in which it was trained, with no or minimal re-training. |
Tasks Fake News Detection, Information Retrieval, Natural Language Inference, Word Embeddings
Published 2018-11-01
URL https://www.aclweb.org/anthology/W18-5528/
PDF https://www.aclweb.org/anthology/W18-5528
PWC https://paperswithcode.com/paper/a-mostly-unlexicalized-model-for-recognizing
Repo
Framework

Word Embedding Approach for Synonym Extraction of Multi-Word Terms

Title Word Embedding Approach for Synonym Extraction of Multi-Word Terms
Authors Amir Hazem, B{'e}atrice Daille
Abstract
Tasks Information Retrieval, Machine Translation, Text Simplification, Word Embeddings, Word Sense Disambiguation
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1045/
PDF https://www.aclweb.org/anthology/L18-1045
PWC https://paperswithcode.com/paper/word-embedding-approach-for-synonym
Repo
Framework

The Lazy Encoder: A Fine-Grained Analysis of the Role of Morphology in Neural Machine Translation

Title The Lazy Encoder: A Fine-Grained Analysis of the Role of Morphology in Neural Machine Translation
Authors Arianna Bisazza, Clara Tump
Abstract Neural sequence-to-sequence models have proven very effective for machine translation, but at the expense of model interpretability. To shed more light into the role played by linguistic structure in the process of neural machine translation, we perform a fine-grained analysis of how various source-side morphological features are captured at different levels of the NMT encoder while varying the target language. Differently from previous work, we find no correlation between the accuracy of source morphology encoding and translation quality. We do find that morphological features are only captured in context and only to the extent that they are directly transferable to the target words.
Tasks Machine Translation, Word Embeddings
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1313/
PDF https://www.aclweb.org/anthology/D18-1313
PWC https://paperswithcode.com/paper/the-lazy-encoder-a-fine-grained-analysis-of
Repo
Framework

Combining Deep Learning and Topic Modeling for Review Understanding in Context-Aware Recommendation

Title Combining Deep Learning and Topic Modeling for Review Understanding in Context-Aware Recommendation
Authors Mingmin Jin, Xin Luo, Huiling Zhu, Hankz Hankui Zhuo
Abstract With the rise of e-commerce, people are accustomed to writing their reviews after receiving the goods. These comments are so important that a bad review can have a direct impact on others buying. Besides, the abundant information within user reviews is very useful for extracting user preferences and item properties. In this paper, we investigate the approach to effectively utilize review information for recommender systems. The proposed model is named LSTM-Topic matrix factorization (LTMF) which integrates both LSTM and Topic Modeling for review understanding. In the experiments on popular review dataset Amazon , our LTMF model outperforms previous proposed HFT model and ConvMF model in rating prediction. Furthermore, LTMF shows the better ability on making topic clustering than traditional topic model based method, which implies integrating the information from deep learning and topic modeling is a meaningful approach to make a better understanding of reviews.
Tasks Recommendation Systems
Published 2018-06-01
URL https://www.aclweb.org/anthology/N18-1145/
PDF https://www.aclweb.org/anthology/N18-1145
PWC https://paperswithcode.com/paper/combining-deep-learning-and-topic-modeling
Repo
Framework

Questionable Answers in Question Answering Research: Reproducibility and Variability of Published Results

Title Questionable Answers in Question Answering Research: Reproducibility and Variability of Published Results
Authors Matt Crane
Abstract {``}Based on theoretical reasoning it has been suggested that the reliability of findings published in the scientific literature decreases with the popularity of a research field{''} (Pfeiffer and Hoffmann, 2009). As we know, deep learning is very popular and the ability to reproduce results is an important part of science. There is growing concern within the deep learning community about the reproducibility of results that are presented. In this paper we present a number of controllable, yet unreported, effects that can substantially change the effectiveness of a sample model, and thusly the reproducibility of those results. Through these environmental effects we show that the commonly held belief that distribution of source code is all that is needed for reproducibility is not enough. Source code without a reproducible environment does not mean anything at all. In addition the range of results produced from these effects can be larger than the majority of incremental improvement reported. |
Tasks Information Retrieval, Question Answering
Published 2018-01-01
URL https://www.aclweb.org/anthology/Q18-1018/
PDF https://www.aclweb.org/anthology/Q18-1018
PWC https://paperswithcode.com/paper/questionable-answers-in-question-answering
Repo
Framework
comments powered by Disqus