January 25, 2020

2611 words 13 mins read

Paper Group NANR 67

Paper Group NANR 67

Reviving a psychometric measure: Classification and prediction of the Operant Motive Test. Coherence models in schizophrenia. Robust Variational Bayesian Point Set Registration. Meta-Semantic Representation for Early Detection of Alzheimer’s Disease. Improving Composition of Sentence Embeddings through the Lens of Statistical Relational Learning. P …

Reviving a psychometric measure: Classification and prediction of the Operant Motive Test

Title Reviving a psychometric measure: Classification and prediction of the Operant Motive Test
Authors Dirk Johann{\ss}en, Chris Biemann, David Scheffer
Abstract Implicit motives allow for the characterization of behavior, subsequent success and long-term development. While this has been operationalized in the operant motive test, research on motives has declined mainly due to labor-intensive and costly human annotation. In this study, we analyze over 200,000 labeled data items from 40,000 participants and utilize them for engineering features for training a logistic model tree machine learning model. It captures manually assigned motives well with an F-score of 80{%}, coming close to the pairwise annotator intraclass correlation coefficient of r = .85. In addition, we found a significant correlation of r = .2 between subsequent academic success and data automatically labeled with our model in an extrinsic evaluation.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-3014/
PDF https://www.aclweb.org/anthology/W19-3014
PWC https://paperswithcode.com/paper/reviving-a-psychometric-measure
Repo
Framework

Coherence models in schizophrenia

Title Coherence models in schizophrenia
Authors S Just, ra, Erik Haegert, Nora Ko{\v{r}}{'a}nov{'a}, Anna-Lena Br{"o}cker, Ivan Nenchev, Jakob Funcke, Christiane Montag, Manfred Stede
Abstract Incoherent discourse in schizophrenia has long been recognized as a dominant symptom of the mental disorder (Bleuler, 1911/1950). Recent studies have used modern sentence and word embeddings to compute coherence metrics for spontaneous speech in schizophrenia. While clinical ratings always have a subjective element, computational linguistic methodology allows quantification of speech abnormalities. Clinical and empirical knowledge from psychiatry provide the theoretical and conceptual basis for modelling. Our study is an interdisciplinary attempt at improving coherence models in schizophrenia. Speech samples were obtained from healthy controls and patients with a diagnosis of schizophrenia or schizoaffective disorder and different severity of positive formal thought disorder. Interviews were transcribed and coherence metrics derived from different embeddings. One model found higher coherence metrics for controls than patients. All other models remained non-significant. More detailed analysis of the data motivates different approaches to improving coherence models in schizophrenia, e.g. by assessing referential abnormalities.
Tasks Word Embeddings
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-3015/
PDF https://www.aclweb.org/anthology/W19-3015
PWC https://paperswithcode.com/paper/coherence-models-in-schizophrenia
Repo
Framework

Robust Variational Bayesian Point Set Registration

Title Robust Variational Bayesian Point Set Registration
Authors Jie Zhou, Xinke Ma, Li Liang, Yang Yang, Shijin Xu, Yuhe Liu, Sim-Heng Ong
Abstract In this work, we propose a hierarchical Bayesian network based point set registration method to solve missing correspondences and various massive outliers. We construct this network first using the finite Student s t latent mixture model (TLMM), in which distributions of latent variables are estimated by a tree-structured variational inference (VI) so that to obtain a tighter lower bound under the Bayesian framework. We then divide the TLMM into two different mixtures with isotropic and anisotropic covariances for correspondences recovering and outliers identification, respectively. Finally, the parameters of mixing proportion and covariances are both taken as latent variables, which benefits explaining of missing correspondences and heteroscedastic outliers. In addition, a cooling schedule is adopted to anneal prior on covariances and scale variables within designed two phases of transformation, it anneal priors on global and local variables to perform a coarse-to- fine registration. In experiments, our method outperforms five state-of-the-art methods in synthetic point set and realistic imaging registrations.
Tasks
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Zhou_Robust_Variational_Bayesian_Point_Set_Registration_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Zhou_Robust_Variational_Bayesian_Point_Set_Registration_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/robust-variational-bayesian-point-set
Repo
Framework

Meta-Semantic Representation for Early Detection of Alzheimer’s Disease

Title Meta-Semantic Representation for Early Detection of Alzheimer’s Disease
Authors Jinho D. Choi, Mengmei Li, Felicia Goldstein, Ihab Hajjar
Abstract This paper presents a new task-oriented meaning representation called meta-semantics, that is designed to detect patients with early symptoms of Alzheimer{'}s disease by analyzing their language beyond a syntactic or semantic level. Meta-semantic representation consists of three parts, entities, predicate argument structures, and discourse attributes, that derive rich knowledge graphs. For this study, 50 controls and 50 patients with mild cognitive impairment (MCI) are selected, and meta-semantic representation is annotated on their speeches transcribed in text. Inter-annotator agreement scores of 88{%}, 82{%}, and 89{%} are achieved for the three types of annotation, respectively. Five analyses are made using this annotation, depicting clear distinctions between the control and MCI groups. Finally, a neural model is trained on features extracted from those analyses to classify MCI patients from normal controls, showing a high accuracy of 82{%} that is very promising.
Tasks Knowledge Graphs
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3309/
PDF https://www.aclweb.org/anthology/W19-3309
PWC https://paperswithcode.com/paper/meta-semantic-representation-for-early
Repo
Framework

Improving Composition of Sentence Embeddings through the Lens of Statistical Relational Learning

Title Improving Composition of Sentence Embeddings through the Lens of Statistical Relational Learning
Authors Damien Sileo, Tim Van de Cruys, Camille Pradel, Philippe Muller
Abstract Various NLP problems – such as the prediction of sentence similarity, entailment, and discourse relations – are all instances of the same general task: the modeling of semantic relations between a pair of textual elements. We call them textual relational problems. A popular model for textual relational problems is to embed sentences into fixed size vectors and use composition functions (e.g. difference or concatenation) of those vectors as features for the prediction. Meanwhile, composition of embeddings has been a main focus within the field of Statistical Relational Learning (SRL) whose goal is to predict relations between entities (typically from knowledge base triples). In this work, we show that textual relational models implicitly use compositions from baseline SRL models. We show that such compositions are not expressive enough for several tasks (e.g. natural language inference). We build on recent SRL models to address textual relational problems, showing that they are more expressive, and can alleviate issues from simpler compositions. The resulting models significantly improve the state of the art in both transferable sentence representation learning and relation prediction.
Tasks Natural Language Inference, Relational Reasoning, Representation Learning, Sentence Embeddings
Published 2019-05-01
URL https://openreview.net/forum?id=SkxZFoAqtQ
PDF https://openreview.net/pdf?id=SkxZFoAqtQ
PWC https://paperswithcode.com/paper/improving-composition-of-sentence-embeddings
Repo
Framework

Predicting Suicide Risk from Online Postings in Reddit The UGent-IDLab submission to the CLPysch 2019 Shared Task A

Title Predicting Suicide Risk from Online Postings in Reddit The UGent-IDLab submission to the CLPysch 2019 Shared Task A
Authors Semere Kiros Bitew, Giannis Bekoulis, Johannes Deleu, Lucas Sterckx, Klim Zaporojets, Thomas Demeester, Chris Develder
Abstract This paper describes IDLab{'}s text classification systems submitted to Task A as part of the CLPsych 2019 shared task. The aim of this shared task was to develop automated systems that predict the degree of suicide risk of people based on their posts on Reddit. Bag-of-words features, emotion features and post level predictions are used to derive user-level predictions. Linear models and ensembles of these models are used to predict final scores. We find that predicting fine-grained risk levels is much more difficult than flagging potentially at-risk users. Furthermore, we do not find clear added value from building richer ensembles compared to simple baselines, given the available training data and the nature of the prediction task.
Tasks Text Classification
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-3019/
PDF https://www.aclweb.org/anthology/W19-3019
PWC https://paperswithcode.com/paper/predicting-suicide-risk-from-online-postings
Repo
Framework

Robustness and Equivariance of Neural Networks

Title Robustness and Equivariance of Neural Networks
Authors Amit Deshpande, Sandesh Kamath, K.V.Subrahmanyam
Abstract Neural networks models are known to be vulnerable to geometric transformations as well as small pixel-wise perturbations of input. Convolutional Neural Networks (CNNs) are translation-equivariant but can be easily fooled using rotations and small pixel-wise perturbations. Moreover, CNNs require sufficient translations in their training data to achieve translation-invariance. Recent work by Cohen & Welling (2016), Worrall et al. (2016), Kondor & Trivedi (2018), Cohen & Welling (2017), Marcos et al. (2017), and Esteves et al. (2018) has gone beyond translations, and constructed rotation-equivariant or more general group-equivariant neural network models. In this paper, we do an extensive empirical study of various rotation-equivariant neural network models to understand how effectively they learn rotations. This includes Group-equivariant Convolutional Networks (GCNNs) by Cohen & Welling (2016), Harmonic Networks (H-Nets) by Worrall et al. (2016), Polar Transformer Networks (PTN) by Esteves et al. (2018) and Rotation equivariant vector field networks by Marcos et al. (2017). We empirically compare the ability of these networks to learn rotations efficiently in terms of their number of parameters, sample complexity, rotation augmentation used in training. We compare them against each other as well as Standard CNNs. We observe that as these rotation-equivariant neural networks learn rotations, they instead become more vulnerable to small pixel-wise adversarial attacks, e.g., Fast Gradient Sign Method (FGSM) and Projected Gradient Descent (PGD), in comparison with Standard CNNs. In other words, robustness to geometric transformations in these models comes at the cost of robustness to small pixel-wise perturbations.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=Ske25sC9FQ
PDF https://openreview.net/pdf?id=Ske25sC9FQ
PWC https://paperswithcode.com/paper/robustness-and-equivariance-of-neural
Repo
Framework

Modelling Uncertainty in Collaborative Document Quality Assessment

Title Modelling Uncertainty in Collaborative Document Quality Assessment
Authors Aili Shen, Daniel Beck, Bahar Salehi, Jianzhong Qi, Timothy Baldwin
Abstract In the context of document quality assessment, previous work has mainly focused on predicting the quality of a document relative to a putative gold standard, without paying attention to the subjectivity of this task. To imitate people{'}s disagreement over inherently subjective tasks such as rating the quality of a Wikipedia article, a document quality assessment system should provide not only a prediction of the article quality but also the uncertainty over its predictions. This motivates us to measure the uncertainty in document quality predictions, in addition to making the label prediction. Experimental results show that both Gaussian processes (GPs) and random forests (RFs) can yield competitive results in predicting the quality of Wikipedia articles, while providing an estimate of uncertainty when there is inconsistency in the quality labels from the Wikipedia contributors. We additionally evaluate our methods in the context of a semi-automated document quality class assignment decision-making process, where there is asymmetric risk associated with overestimates and underestimates of document quality. Our experiments suggest that GPs provide more reliable estimates in this context.
Tasks Decision Making, Gaussian Processes
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5525/
PDF https://www.aclweb.org/anthology/D19-5525
PWC https://paperswithcode.com/paper/modelling-uncertainty-in-collaborative
Repo
Framework

Suicide Risk Assessment on Social Media: USI-UPF at the CLPsych 2019 Shared Task

Title Suicide Risk Assessment on Social Media: USI-UPF at the CLPsych 2019 Shared Task
Authors Esteban R{'\i}ssola, Diana Ram{'\i}rez-Cifuentes, Ana Freire, Fabio Crestani
Abstract This paper describes the participation of the USI-UPF team at the shared task of the 2019 Computational Linguistics and Clinical Psychology Workshop (CLPsych2019). The goal is to assess the degree of suicide risk of social media users given a labelled dataset with their posts. An appropriate suicide risk assessment, with the usage of automated methods, can assist experts on the detection of people at risk and eventually contribute to prevent suicide. We propose a set of machine learning models with features based on lexicons, word embeddings, word level n-grams, and statistics extracted from users{'} posts. The results show that the most effective models for the tasks are obtained integrating lexicon-based features, a selected set of n-grams, and statistical measures.
Tasks Word Embeddings
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-3021/
PDF https://www.aclweb.org/anthology/W19-3021
PWC https://paperswithcode.com/paper/suicide-risk-assessment-on-social-media-usi
Repo
Framework

A Survey of Recent Advances in Efficient Parsing for Graph Grammars

Title A Survey of Recent Advances in Efficient Parsing for Graph Grammars
Authors Frank Drewes
Abstract
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-3102/
PDF https://www.aclweb.org/anthology/W19-3102
PWC https://paperswithcode.com/paper/a-survey-of-recent-advances-in-efficient
Repo
Framework

Latent Variable Grammars for Discontinuous Parsing

Title Latent Variable Grammars for Discontinuous Parsing
Authors Kilian Gebhardt
Abstract
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-3103/
PDF https://www.aclweb.org/anthology/W19-3103
PWC https://paperswithcode.com/paper/latent-variable-grammars-for-discontinuous
Repo
Framework

Composition of Sentence Embeddings: Lessons from Statistical Relational Learning

Title Composition of Sentence Embeddings: Lessons from Statistical Relational Learning
Authors Damien Sileo, Tim Van De Cruys, Camille Pradel, Philippe Muller
Abstract Various NLP problems {–} such as the prediction of sentence similarity, entailment, and discourse relations {–} are all instances of the same general task: the modeling of semantic relations between a pair of textual elements. A popular model for such problems is to embed sentences into fixed size vectors, and use composition functions (e.g. concatenation or sum) of those vectors as features for the prediction. At the same time, composition of embeddings has been a main focus within the field of Statistical Relational Learning (SRL) whose goal is to predict relations between entities (typically from knowledge base triples). In this article, we show that previous work on relation prediction between texts implicitly uses compositions from baseline SRL models. We show that such compositions are not expressive enough for several tasks (e.g. natural language inference). We build on recent SRL models to address textual relational problems, showing that they are more expressive, and can alleviate issues from simpler compositions. The resulting models significantly improve the state of the art in both transferable sentence representation learning and relation prediction.
Tasks Natural Language Inference, Relational Reasoning, Representation Learning, Sentence Embeddings
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-1004/
PDF https://www.aclweb.org/anthology/S19-1004
PWC https://paperswithcode.com/paper/composition-of-embeddings-lessons-from
Repo
Framework

Bottom-Up Unranked Tree-to-Graph Transducers for Translation into Semantic Graphs

Title Bottom-Up Unranked Tree-to-Graph Transducers for Translation into Semantic Graphs
Authors Johanna Bj{"o}rklund, Shay B. Cohen, Frank Drewes, Giorgio Satta
Abstract We propose a formal model for translating unranked syntactic trees, such as dependency trees, into semantic graphs. These tree-to-graph transducers can serve as a formal basis of transition systems for semantic parsing which recently have been shown to perform very well, yet hitherto lack formalization. Our model features {``}extended{''} rules and an arc-factored normal form, comes with an efficient translation algorithm, and can be equipped with weights in a straightforward manner. |
Tasks Semantic Parsing
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-3104/
PDF https://www.aclweb.org/anthology/W19-3104
PWC https://paperswithcode.com/paper/bottom-up-unranked-tree-to-graph-transducers
Repo
Framework

TOI-CNN: a Solution of Information Extraction on Chinese Insurance Policy

Title TOI-CNN: a Solution of Information Extraction on Chinese Insurance Policy
Authors Lin Sun, Kai Zhang, Fule Ji, Zhenhua Yang
Abstract Contract analysis can significantly ease the work for humans using AI techniques. This paper shows a problem of Element Tagging on Insurance Policy (ETIP). A novel Text-Of-Interest Convolutional Neural Network (TOI-CNN) is proposed for the ETIP solution. We introduce a TOI pooling layer to replace traditional pooling layer for processing the nested phrasal or clausal elements in insurance policies. The advantage of TOI pooling layer is that the nested elements from one sentence could share computation and context in the forward and backward passes. The computation of backpropagation through TOI pooling is also demonstrated in the paper. We have collected a large Chinese insurance contract dataset and labeled the critical elements of seven categories to test the performance of the proposed method. The results show the promising performance of our method in the ETIP problem.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-2022/
PDF https://www.aclweb.org/anthology/N19-2022
PWC https://paperswithcode.com/paper/toi-cnn-a-solution-of-information-extraction
Repo
Framework

On the Compression of Lexicon Transducers

Title On the Compression of Lexicon Transducers
Authors Marco Cognetta, Cyril Allauzen, Michael Riley
Abstract In finite-state language processing pipelines, a lexicon is often a key component. It needs to be comprehensive to ensure accuracy, reducing out-of-vocabulary misses. However, in memory-constrained environments (e.g., mobile phones), the size of the component automata must be kept small. Indeed, a delicate balance between comprehensiveness, speed, and memory must be struck to conform to device requirements while providing a good user experience.In this paper, we describe a compression scheme for lexicons when represented as finite-state transducers. We efficiently encode the graph of the transducer while storing transition labels separately. The graph encoding scheme is based on the LOUDS (Level Order Unary Degree Sequence) tree representation, which has constant time tree traversal for queries while being information-theoretically optimal in space. We find that our encoding is near the theoretical lower bound for such graphs and substantially outperforms more traditional representations in space while remaining competitive in latency benchmarks.
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-3105/
PDF https://www.aclweb.org/anthology/W19-3105
PWC https://paperswithcode.com/paper/on-the-compression-of-lexicon-transducers
Repo
Framework
comments powered by Disqus