July 26, 2019

2537 words 12 mins read

Paper Group NANR 84

Paper Group NANR 84

Proceedings of the IJCNLP 2017, Shared Tasks. Predicting Users’ Negative Feedbacks in Multi-Turn Human-Computer Dialogues. Multi-Grained Chinese Word Segmentation. Neural Joint Model for Transition-based Chinese Syntactic Analysis. OPI-JSA at SemEval-2017 Task 1: Application of Ensemble learning for computing semantic textual similarity. Semantic W …

Proceedings of the IJCNLP 2017, Shared Tasks

Title Proceedings of the IJCNLP 2017, Shared Tasks
Authors
Abstract
Tasks
Published 2017-12-01
URL https://www.aclweb.org/anthology/I17-4000/
PDF https://www.aclweb.org/anthology/I17-4000
PWC https://paperswithcode.com/paper/proceedings-of-the-ijcnlp-2017-shared-tasks
Repo
Framework

Predicting Users’ Negative Feedbacks in Multi-Turn Human-Computer Dialogues

Title Predicting Users’ Negative Feedbacks in Multi-Turn Human-Computer Dialogues
Authors Xin Wang, Jianan Wang, Yuanchao Liu, Xiaolong Wang, Zhuoran Wang, Baoxun Wang
Abstract User experience is essential for human-computer dialogue systems. However, it is impractical to ask users to provide explicit feedbacks when the agents{'} responses displease them. Therefore, in this paper, we explore to predict users{'} imminent dissatisfactions caused by intelligent agents by analysing the existing utterances in the dialogue sessions. To our knowledge, this is the first work focusing on this task. Several possible factors that trigger negative emotions are modelled. A relation sequence model (RSM) is proposed to encode the sequence of appropriateness of current response with respect to the earlier utterances. The experimental results show that the proposed structure is effective in modelling emotional risk (possibility of negative feedback) than existing conversation modelling approaches. Besides, strategies of obtaining distance supervision data for pre-training are also discussed in this work. Balanced sampling with respect to the last response in the distance supervision data are shown to be reliable for data augmentation.
Tasks Data Augmentation
Published 2017-11-01
URL https://www.aclweb.org/anthology/I17-1072/
PDF https://www.aclweb.org/anthology/I17-1072
PWC https://paperswithcode.com/paper/predicting-users-negative-feedbacks-in-multi
Repo
Framework

Multi-Grained Chinese Word Segmentation

Title Multi-Grained Chinese Word Segmentation
Authors Chen Gong, Zhenghua Li, Min Zhang, Xinzhou Jiang
Abstract Traditionally, word segmentation (WS) adopts the single-grained formalism, where a sentence corresponds to a single word sequence. However, Sproat et al. (1997) show that the inter-native-speaker consistency ratio over Chinese word boundaries is only 76{%}, indicating single-grained WS (SWS) imposes unnecessary challenges on both manual annotation and statistical modeling. Moreover, WS results of different granularities can be complementary and beneficial for high-level applications. This work proposes and addresses multi-grained WS (MWS). We build a large-scale pseudo MWS dataset for model training and tuning by leveraging the annotation heterogeneity of three SWS datasets. Then we manually annotate 1,500 test sentences with true MWS annotations. Finally, we propose three benchmark approaches by casting MWS as constituent parsing and sequence labeling. Experiments and analysis lead to many interesting findings.
Tasks Chinese Word Segmentation, Language Modelling
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1072/
PDF https://www.aclweb.org/anthology/D17-1072
PWC https://paperswithcode.com/paper/multi-grained-chinese-word-segmentation
Repo
Framework

Neural Joint Model for Transition-based Chinese Syntactic Analysis

Title Neural Joint Model for Transition-based Chinese Syntactic Analysis
Authors Shuhei Kurita, Daisuke Kawahara, Sadao Kurohashi
Abstract We present neural network-based joint models for Chinese word segmentation, POS tagging and dependency parsing. Our models are the first neural approaches for fully joint Chinese analysis that is known to prevent the error propagation problem of pipeline models. Although word embeddings play a key role in dependency parsing, they cannot be applied directly to the joint task in the previous work. To address this problem, we propose embeddings of character strings, in addition to words. Experiments show that our models outperform existing systems in Chinese word segmentation and POS tagging, and perform preferable accuracies in dependency parsing. We also explore bi-LSTM models with fewer features.
Tasks Chinese Word Segmentation, Dependency Parsing, Word Embeddings
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-1111/
PDF https://www.aclweb.org/anthology/P17-1111
PWC https://paperswithcode.com/paper/neural-joint-model-for-transition-based
Repo
Framework

OPI-JSA at SemEval-2017 Task 1: Application of Ensemble learning for computing semantic textual similarity

Title OPI-JSA at SemEval-2017 Task 1: Application of Ensemble learning for computing semantic textual similarity
Authors Martyna {'S}piewak, Piotr Sobecki, Daniel Kara{'s}
Abstract Semantic Textual Similarity (STS) evaluation assesses the degree to which two parts of texts are similar, based on their semantic evaluation. In this paper, we describe three models submitted to STS SemEval 2017. Given two English parts of a text, each of proposed methods outputs the assessment of their semantic similarity. We propose an approach for computing monolingual semantic textual similarity based on an ensemble of three distinct methods. Our model consists of recursive neural network (RNN) text auto-encoders ensemble with supervised a model of vectorized sentences using reduced part of speech (PoS) weighted word embeddings as well as unsupervised a method based on word coverage (TakeLab). Additionally, we enrich our model with additional features that allow disambiguation of ensemble methods based on their efficiency. We have used Multi-Layer Perceptron as an ensemble classifier basing on estimations of trained Gradient Boosting Regressors. Results of our research proves that using such ensemble leads to a higher accuracy due to a fact that each member-algorithm tends to specialize in particular type of sentences. Simple model based on PoS weighted Word2Vec word embeddings seem to improve performance of more complex RNN based auto-encoders in the ensemble. In the monolingual English-English STS subtask our Ensemble based model achieved mean Pearson correlation of .785 compared with human annotators.
Tasks Semantic Similarity, Semantic Textual Similarity, Word Embeddings
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2018/
PDF https://www.aclweb.org/anthology/S17-2018
PWC https://paperswithcode.com/paper/opi-jsa-at-semeval-2017-task-1-application-of
Repo
Framework

Semantic Word Clusters Using Signed Spectral Clustering

Title Semantic Word Clusters Using Signed Spectral Clustering
Authors Jo{~a}o Sedoc, Jean Gallier, Dean Foster, Lyle Ungar
Abstract Vector space representations of words capture many aspects of word similarity, but such methods tend to produce vector spaces in which antonyms (as well as synonyms) are close to each other. For spectral clustering using such word embeddings, words are points in a vector space where synonyms are linked with positive weights, while antonyms are linked with negative weights. We present a new signed spectral normalized graph cut algorithm, \textit{signed clustering}, that overlays existing thesauri upon distributionally derived vector representations of words, so that antonym relationships between word pairs are represented by negative weights. Our signed clustering algorithm produces clusters of words that simultaneously capture distributional and synonym relations. By using randomized spectral decomposition (Halko et al., 2011) and sparse matrices, our method is both fast and scalable. We validate our clusters using datasets containing human judgments of word pair similarities and show the benefit of using our word clusters for sentiment prediction.
Tasks Graph Clustering, Semantic Textual Similarity, Word Embeddings
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-1087/
PDF https://www.aclweb.org/anthology/P17-1087
PWC https://paperswithcode.com/paper/semantic-word-clusters-using-signed-spectral
Repo
Framework

Multimodal Machine Learning: Integrating Language, Vision and Speech

Title Multimodal Machine Learning: Integrating Language, Vision and Speech
Authors Louis-Philippe Morency, Tadas Baltru{\v{s}}aitis
Abstract Multimodal machine learning is a vibrant multi-disciplinary research field which addresses some of the original goals of artificial intelligence by integrating and modeling multiple communicative modalities, including linguistic, acoustic and visual messages. With the initial research on audio-visual speech recognition and more recently with image and video captioning projects, this research field brings some unique challenges for multimodal researchers given the heterogeneity of the data and the contingency often found between modalities.This tutorial builds upon a recent course taught at Carnegie Mellon University during the Spring 2016 semester (CMU course 11-777) and two tutorials presented at CVPR 2016 and ICMI 2016. The present tutorial will review fundamental concepts of machine learning and deep neural networks before describing the five main challenges in multimodal machine learning: (1) multimodal representation learning, (2) translation {&} mapping, (3) modality alignment, (4) multimodal fusion and (5) co-learning. The tutorial will also present state-of-the-art algorithms that were recently proposed to solve multimodal applications such as image captioning, video descriptions and visual question-answer. We will also discuss the current and upcoming challenges.
Tasks Audio-Visual Speech Recognition, Image Captioning, Question Answering, Representation Learning, Speech Recognition, Video Captioning, Visual Question Answering, Visual Speech Recognition, Zero-Shot Learning
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-5002/
PDF https://www.aclweb.org/anthology/P17-5002
PWC https://paperswithcode.com/paper/multimodal-machine-learning-integrating
Repo
Framework

Ranking Kernels for Structures and Embeddings: A Hybrid Preference and Classification Model

Title Ranking Kernels for Structures and Embeddings: A Hybrid Preference and Classification Model
Authors Kateryna Tymoshenko, Daniele Bonadiman, Aless Moschitti, ro
Abstract Recent work has shown that Tree Kernels (TKs) and Convolutional Neural Networks (CNNs) obtain the state of the art in answer sentence reranking. Additionally, their combination used in Support Vector Machines (SVMs) is promising as it can exploit both the syntactic patterns captured by TKs and the embeddings learned by CNNs. However, the embeddings are constructed according to a classification function, which is not directly exploitable in the preference ranking algorithm of SVMs. In this work, we propose a new hybrid approach combining preference ranking applied to TKs and pointwise ranking applied to CNNs. We show that our approach produces better results on two well-known and rather different datasets: WikiQA for answer sentence selection and SemEval cQA for comment selection in Community Question Answering.
Tasks Community Question Answering, Learning-To-Rank, Question Answering, Word Embeddings
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1093/
PDF https://www.aclweb.org/anthology/D17-1093
PWC https://paperswithcode.com/paper/ranking-kernels-for-structures-and-embeddings
Repo
Framework

ECNU at SemEval-2017 Task 8: Rumour Evaluation Using Effective Features and Supervised Ensemble Models

Title ECNU at SemEval-2017 Task 8: Rumour Evaluation Using Effective Features and Supervised Ensemble Models
Authors Feixiang Wang, Man Lan, Yuanbin Wu
Abstract This paper describes our submissions to task 8 in SemEval 2017, i.e., Determining rumour veracity and support for rumours. Given a rumoured tweet and a lot of reply tweets, the subtask A is to label whether these tweets are support, deny, query or comment, and the subtask B aims to predict the veracity (i.e., true, false, and unverified) with a confidence (in range of 0-1) of the given rumoured tweet. For both subtasks, we adopted supervised machine learning methods, incorporating rich features. Since training data is imbalanced, we specifically designed a two-step classifier to address subtask A .
Tasks Rumour Detection, Sentiment Analysis
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2086/
PDF https://www.aclweb.org/anthology/S17-2086
PWC https://paperswithcode.com/paper/ecnu-at-semeval-2017-task-8-rumour-evaluation
Repo
Framework

Dependency Parsing with Partial Annotations: An Empirical Comparison

Title Dependency Parsing with Partial Annotations: An Empirical Comparison
Authors Yue Zhang, Zhenghua Li, Jun Lang, Qingrong Xia, Min Zhang
Abstract This paper describes and compares two straightforward approaches for dependency parsing with partial annotations (PA). The first approach is based on a forest-based training objective for two CRF parsers, i.e., a biaffine neural network graph-based parser (Biaffine) and a traditional log-linear graph-based parser (LLGPar). The second approach is based on the idea of constrained decoding for three parsers, i.e., a traditional linear graph-based parser (LGPar), a globally normalized neural network transition-based parser (GN3Par) and a traditional linear transition-based parser (LTPar). For the test phase, constrained decoding is also used for completing partial trees. We conduct experiments on Penn Treebank under three different settings for simulating PA, i.e., random, most uncertain, and divergent outputs from the five parsers. The results show that LLGPar is most effective in directly learning from PA, and other parsers can achieve best performance when PAs are completed into full trees by LLGPar.
Tasks Active Learning, Dependency Parsing
Published 2017-11-01
URL https://www.aclweb.org/anthology/I17-1006/
PDF https://www.aclweb.org/anthology/I17-1006
PWC https://paperswithcode.com/paper/dependency-parsing-with-partial-annotations
Repo
Framework

Towards Lower Bounds on Number of Dimensions for Word Embeddings

Title Towards Lower Bounds on Number of Dimensions for Word Embeddings
Authors Kevin Patel, Pushpak Bhattacharyya
Abstract Word embeddings are a relatively new addition to the modern NLP researcher{'}s toolkit. However, unlike other tools, word embeddings are used in a black box manner. There are very few studies regarding various hyperparameters. One such hyperparameter is the dimension of word embeddings. They are rather decided based on a rule of thumb: in the range 50 to 300. In this paper, we show that the dimension should instead be chosen based on corpus statistics. More specifically, we show that the number of pairwise equidistant words of the corpus vocabulary (as defined by some distance/similarity metric) gives a lower bound on the the number of dimensions , and going below this bound results in degradation of quality of learned word embeddings. Through our evaluations on standard word embedding evaluation tasks, we show that for dimensions higher than or equal to the bound, we get better results as compared to the ones below it.
Tasks Named Entity Recognition, Part-Of-Speech Tagging, Sarcasm Detection, Sentence Classification, Sentiment Analysis, Word Embeddings
Published 2017-11-01
URL https://www.aclweb.org/anthology/I17-2006/
PDF https://www.aclweb.org/anthology/I17-2006
PWC https://paperswithcode.com/paper/towards-lower-bounds-on-number-of-dimensions
Repo
Framework

Learning to See Physics via Visual De-animation

Title Learning to See Physics via Visual De-animation
Authors Jiajun Wu, Erika Lu, Pushmeet Kohli, Bill Freeman, Josh Tenenbaum
Abstract We introduce a paradigm for understanding physical scenes without human annotations. At the core of our system is a physical world representation that is first recovered by a perception module and then utilized by physics and graphics engines. During training, the perception module and the generative models learn by visual de-animation — interpreting and reconstructing the visual information stream. During testing, the system first recovers the physical world state, and then uses the generative models for reasoning and future prediction. Even more so than forward simulation, inverting a physics or graphics engine is a computationally hard problem; we overcome this challenge by using a convolutional inversion network. Our system quickly recognizes the physical world state from appearance and motion cues, and has the flexibility to incorporate both differentiable and non-differentiable physics and graphics engines. We evaluate our system on both synthetic and real datasets involving multiple physical scenes, and demonstrate that our system performs well on both physical state estimation and reasoning problems. We further show that the knowledge learned on the synthetic dataset generalizes to constrained real images.
Tasks Future prediction
Published 2017-12-01
URL http://papers.nips.cc/paper/6620-learning-to-see-physics-via-visual-de-animation
PDF http://papers.nips.cc/paper/6620-learning-to-see-physics-via-visual-de-animation.pdf
PWC https://paperswithcode.com/paper/learning-to-see-physics-via-visual-de
Repo
Framework

Benben: A Chinese Intelligent Conversational Robot

Title Benben: A Chinese Intelligent Conversational Robot
Authors Wei-Nan Zhang, Ting Liu, Bing Qin, Yu Zhang, Wanxiang Che, Yanyan Zhao, Xiao Ding
Abstract
Tasks Chinese Word Segmentation, Dependency Parsing, Information Retrieval, Intent Detection, Named Entity Recognition, Part-Of-Speech Tagging, Question Answering, Reading Comprehension, Sentiment Analysis, Word Sense Disambiguation
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-4003/
PDF https://www.aclweb.org/anthology/P17-4003
PWC https://paperswithcode.com/paper/benben-a-chinese-intelligent-conversational
Repo
Framework

Textually Summarising Incomplete Data

Title Textually Summarising Incomplete Data
Authors Stephanie Inglis, Ehud Reiter, Somayajulu Sripada
Abstract Many data-to-text NLG systems work with data sets which are incomplete, ie some of the data is missing. We have worked with data journalists to understand how they describe incomplete data, and are building NLG algorithms based on these insights. A pilot evaluation showed mixed results, and highlighted several areas where we need to improve our system.
Tasks Text Generation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-3535/
PDF https://www.aclweb.org/anthology/W17-3535
PWC https://paperswithcode.com/paper/textually-summarising-incomplete-data
Repo
Framework

Data-Driven Broad-Coverage Grammars for Opinionated Natural Language Generation (ONLG)

Title Data-Driven Broad-Coverage Grammars for Opinionated Natural Language Generation (ONLG)
Authors Tomer Cagan, Stefan L. Frank, Reut Tsarfaty
Abstract Opinionated Natural Language Generation (ONLG) is a new, challenging, task that aims to automatically generate human-like, subjective, responses to opinionated articles online. We present a data-driven architecture for ONLG that generates subjective responses triggered by users{'} agendas, consisting of topics and sentiments, and based on wide-coverage automatically-acquired generative grammars. We compare three types of grammatical representations that we design for ONLG, which interleave different layers of linguistic information and are induced from a new, enriched dataset we developed. Our evaluation shows that generation with Relational-Realizational (Tsarfaty and Sima{'}an, 2008) inspired grammar gets better language model scores than lexicalized grammars {`}a la Collins (2003), and that the latter gets better human-evaluation scores. We also show that conditioning the generation on topic models makes generated responses more relevant to the document content. |
Tasks Language Modelling, Text Generation, Topic Models
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-1122/
PDF https://www.aclweb.org/anthology/P17-1122
PWC https://paperswithcode.com/paper/data-driven-broad-coverage-grammars-for
Repo
Framework
comments powered by Disqus