July 26, 2019

1782 words 9 mins read

Paper Group NANR 149

Paper Group NANR 149

Proceedings of the 1st Workshop on Explainable Computational Intelligence (XCI 2017). A Simple Method for Clarifying Sentences with Coordination Ambiguities. Partitioned Tensor Factorizations for Learning Mixed Membership Models. Learning Infinite Layer Networks without the Kernel Trick. Detecting Personal Medication Intake in Twitter: An Annotated …

Proceedings of the 1st Workshop on Explainable Computational Intelligence (XCI 2017)

Title Proceedings of the 1st Workshop on Explainable Computational Intelligence (XCI 2017)
Authors
Abstract
Tasks
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-3700/
PDF https://www.aclweb.org/anthology/W17-3700
PWC https://paperswithcode.com/paper/proceedings-of-the-1st-workshop-on-1
Repo
Framework

A Simple Method for Clarifying Sentences with Coordination Ambiguities

Title A Simple Method for Clarifying Sentences with Coordination Ambiguities
Authors Michael White, Manjuan Duan, David L. King
Abstract
Tasks
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-3702/
PDF https://www.aclweb.org/anthology/W17-3702
PWC https://paperswithcode.com/paper/a-simple-method-for-clarifying-sentences-with
Repo
Framework

Partitioned Tensor Factorizations for Learning Mixed Membership Models

Title Partitioned Tensor Factorizations for Learning Mixed Membership Models
Authors Zilong Tan, Sayan Mukherjee
Abstract We present an efficient algorithm for learning mixed membership models when the number of variables p is much larger than the number of hidden components k. This algorithm reduces the computational complexity of state-of-the-art tensor methods, which require decomposing an $O(p^3)$ tensor, to factorizing $O(p/k)$ sub-tensors each of size $O(k^3)$. In addition, we address the issue of negative entries in the empirical method of moments based estimators. We provide sufficient conditions under which our approach has provable guarantees. Our approach obtains competitive empirical results on both simulated and real data.
Tasks
Published 2017-08-01
URL https://icml.cc/Conferences/2017/Schedule?showEvent=551
PDF http://proceedings.mlr.press/v70/tan17a/tan17a.pdf
PWC https://paperswithcode.com/paper/partitioned-tensor-factorizations-for
Repo
Framework

Learning Infinite Layer Networks without the Kernel Trick

Title Learning Infinite Layer Networks without the Kernel Trick
Authors Roi Livni, Daniel Carmon, Amir Globerson
Abstract Infinite Layer Networks (ILN) have been proposed as an architecture that mimics neural networks while enjoying some of the advantages of kernel methods. ILN are networks that integrate over infinitely many nodes within a single hidden layer. It has been demonstrated by several authors that the problem of learning ILN can be reduced to the kernel trick, implying that whenever a certain integral can be computed analytically they are efficiently learnable. In this work we give an online algorithm for ILN, which avoids the kernel trick assumption. More generally and of independent interest, we show that kernel methods in general can be exploited even when the kernel cannot be efficiently computed but can only be estimated via sampling. We provide a regret analysis for our algorithm, showing that it matches the sample complexity of methods which have access to kernel values. Thus, our method is the first to demonstrate that the kernel trick is not necessary, as such, and random features suffice to obtain comparable performance.
Tasks
Published 2017-08-01
URL https://icml.cc/Conferences/2017/Schedule?showEvent=755
PDF http://proceedings.mlr.press/v70/livni17a/livni17a.pdf
PWC https://paperswithcode.com/paper/learning-infinite-layer-networks-without-the
Repo
Framework

Detecting Personal Medication Intake in Twitter: An Annotated Corpus and Baseline Classification System

Title Detecting Personal Medication Intake in Twitter: An Annotated Corpus and Baseline Classification System
Authors Ari Klein, Abeed Sarker, Masoud Rouhizadeh, Karen O{'}Connor, Graciela Gonzalez
Abstract Social media sites (e.g., Twitter) have been used for surveillance of drug safety at the population level, but studies that focus on the effects of medications on specific sets of individuals have had to rely on other sources of data. Mining social media data for this in-formation would require the ability to distinguish indications of personal medication in-take in this media. Towards that end, this paper presents an annotated corpus that can be used to train machine learning systems to determine whether a tweet that mentions a medication indicates that the individual posting has taken that medication at a specific time. To demonstrate the utility of the corpus as a training set, we present baseline results of supervised classification.
Tasks Epidemiology
Published 2017-08-01
URL https://www.aclweb.org/anthology/W17-2316/
PDF https://www.aclweb.org/anthology/W17-2316
PWC https://paperswithcode.com/paper/detecting-personal-medication-intake-in
Repo
Framework

Proceedings of the Second Conference on Machine Translation

Title Proceedings of the Second Conference on Machine Translation
Authors
Abstract
Tasks Machine Translation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4700/
PDF https://www.aclweb.org/anthology/W17-4700
PWC https://paperswithcode.com/paper/proceedings-of-the-second-conference-on
Repo
Framework

Evaluation Metrics for Machine Reading Comprehension: Prerequisite Skills and Readability

Title Evaluation Metrics for Machine Reading Comprehension: Prerequisite Skills and Readability
Authors Saku Sugawara, Yusuke Kido, Hikaru Yokono, Akiko Aizawa
Abstract Knowing the quality of reading comprehension (RC) datasets is important for the development of natural-language understanding systems. In this study, two classes of metrics were adopted for evaluating RC datasets: prerequisite skills and readability. We applied these classes to six existing datasets, including MCTest and SQuAD, and highlighted the characteristics of the datasets according to each metric and the correlation between the two classes. Our dataset analysis suggests that the readability of RC datasets does not directly affect the question difficulty and that it is possible to create an RC dataset that is easy to read but difficult to answer.
Tasks Coreference Resolution, Machine Reading Comprehension, Reading Comprehension
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-1075/
PDF https://www.aclweb.org/anthology/P17-1075
PWC https://paperswithcode.com/paper/evaluation-metrics-for-machine-reading
Repo
Framework

Generating Text with Correct Verb Conjugation: Proposal for a New Automatic Conjugator with NooJ

Title Generating Text with Correct Verb Conjugation: Proposal for a New Automatic Conjugator with NooJ
Authors H{'e}la Fehri, Sondes Dardour
Abstract
Tasks Text Generation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-3806/
PDF https://www.aclweb.org/anthology/W17-3806
PWC https://paperswithcode.com/paper/generating-text-with-correct-verb-conjugation
Repo
Framework

Dictionary Learning Based on Sparse Distribution Tomography

Title Dictionary Learning Based on Sparse Distribution Tomography
Authors Pedram Pad, Farnood Salehi, Elisa Celis, Patrick Thiran, Michael Unser
Abstract We propose a new statistical dictionary learning algorithm for sparse signals that is based on an $\alpha$-stable innovation model. The parameters of the underlying model—that is, the atoms of the dictionary, the sparsity index $\alpha$ and the dispersion of the transform-domain coefficients—are recovered using a new type of probability distribution tomography. Specifically, we drive our estimator with a series of random projections of the data, which results in an efficient algorithm. Moreover, since the projections are achieved using linear combinations, we can invoke the generalized central limit theorem to justify the use of our method for sparse signals that are not necessarily $\alpha$-stable. We evaluate our algorithm by performing two types of experiments: image in-painting and image denoising. In both cases, we find that our approach is competitive with state-of-the-art dictionary learning techniques. Beyond the algorithm itself, two aspects of this study are interesting in their own right. The first is our statistical formulation of the problem, which unifies the topics of dictionary learning and independent component analysis. The second is a generalization of a classical theorem about isometries of $\ell_p$-norms that constitutes the foundation of our approach.
Tasks Denoising, Dictionary Learning, Image Denoising
Published 2017-08-01
URL https://icml.cc/Conferences/2017/Schedule?showEvent=662
PDF http://proceedings.mlr.press/v70/pad17a/pad17a.pdf
PWC https://paperswithcode.com/paper/dictionary-learning-based-on-sparse
Repo
Framework

An Essay on Self-explanatory Computational Intelligence: A Linguistic Model of Data Processing Systems

Title An Essay on Self-explanatory Computational Intelligence: A Linguistic Model of Data Processing Systems
Authors Jose M. Alonso, Gracian Trivino
Abstract
Tasks Text Generation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-3704/
PDF https://www.aclweb.org/anthology/W17-3704
PWC https://paperswithcode.com/paper/an-essay-on-self-explanatory-computational
Repo
Framework

Evaluating an Automata Approach to Query Containment

Title Evaluating an Automata Approach to Query Containment
Authors Michael Minock
Abstract
Tasks Automated Theorem Proving, Text Generation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4010/
PDF https://www.aclweb.org/anthology/W17-4010
PWC https://paperswithcode.com/paper/evaluating-an-automata-approach-to-query
Repo
Framework

Poet’s Little Helper: A methodology for computer-based poetry generation. A case study for the Basque language

Title Poet’s Little Helper: A methodology for computer-based poetry generation. A case study for the Basque language
Authors Aitzol Astigarraga, Jos{'e} Mar{'\i}a Mart{'\i}nez-Otzeta, Igor Rodriguez, Basilio Sierra, Elena Lazkano
Abstract
Tasks
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-3901/
PDF https://www.aclweb.org/anthology/W17-3901
PWC https://paperswithcode.com/paper/poets-little-helper-a-methodology-for
Repo
Framework

Stochastic DCA for the Large-sum of Non-convex Functions Problem and its Application to Group Variable Selection in Classification

Title Stochastic DCA for the Large-sum of Non-convex Functions Problem and its Application to Group Variable Selection in Classification
Authors Hoai An Le Thi, Hoai Minh Le, Duy Nhat Phan, Bach Tran
Abstract In this paper, we present a stochastic version of DCA (Difference of Convex functions Algorithm) to solve a class of optimization problems whose objective function is a large sum of non-convex functions and a regularization term. We consider the $\ell_{2,0}$ regularization to deal with the group variables selection. By exploiting the special structure of the problem, we propose an efficient DC decomposition for which the corresponding stochastic DCA scheme is very inexpensive: it only requires the projection of points onto balls that is explicitly computed. As an application, we applied our algorithm for the group variables selection in multiclass logistic regression. Numerical experiments on several benchmark datasets and synthetic datasets illustrate the efficiency of our algorithm and its superiority over well-known methods, with respect to classification accuracy, sparsity of solution as well as running time.
Tasks
Published 2017-08-01
URL https://icml.cc/Conferences/2017/Schedule?showEvent=886
PDF http://proceedings.mlr.press/v70/thi17a/thi17a.pdf
PWC https://paperswithcode.com/paper/stochastic-dca-for-the-large-sum-of-non
Repo
Framework

Building Lexical Vector Representations from Concept Definitions

Title Building Lexical Vector Representations from Concept Definitions
Authors Danilo Silva de Carvalho, Minh Le Nguyen
Abstract The use of distributional language representations have opened new paths in solving a variety of NLP problems. However, alternative approaches can take advantage of information unavailable through pure statistical means. This paper presents a method for building vector representations from meaning unit blocks called concept definitions, which are obtained by extracting information from a curated linguistic resource (Wiktionary). The representations obtained in this way can be compared through conventional cosine similarity and are also interpretable by humans. Evaluation was conducted in semantic similarity and relatedness test sets, with results indicating a performance comparable to other methods based on single linguistic resource extraction. The results also indicate noticeable performance gains when combining distributional similarity scores with the ones obtained using this approach. Additionally, a discussion on the proposed method{'}s shortcomings is provided in the analysis of error cases.
Tasks Dependency Parsing, Machine Translation, Named Entity Recognition, Semantic Similarity, Semantic Textual Similarity
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-1085/
PDF https://www.aclweb.org/anthology/E17-1085
PWC https://paperswithcode.com/paper/building-lexical-vector-representations-from
Repo
Framework

UINSUSKA-TiTech at SemEval-2017 Task 3: Exploiting Word Importance Levels for Similarity Features for CQA

Title UINSUSKA-TiTech at SemEval-2017 Task 3: Exploiting Word Importance Levels for Similarity Features for CQA
Authors Surya Agustian, Hiroya Takamura
Abstract The majority of core techniques to solve many problems in Community Question Answering (CQA) task rely on similarity computation. This work focuses on similarity between two sentences (or questions in subtask B) based on word embeddings. We exploit words importance levels in sentences or questions for similarity features, for classification and ranking with machine learning. Using only 2 types of similarity metric, our proposed method has shown comparable results with other complex systems. This method on subtask B 2017 dataset is ranked on position 7 out of 13 participants. Evaluation on 2016 dataset is on position 8 of 12, outperforms some complex systems. Further, this finding is explorable and potential to be used as baseline and extensible for many tasks in CQA and other textual similarity based system.
Tasks Community Question Answering, Knowledge Graphs, Question Answering, Semantic Textual Similarity, Word Embeddings
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2061/
PDF https://www.aclweb.org/anthology/S17-2061
PWC https://paperswithcode.com/paper/uinsuska-titech-at-semeval-2017-task-3
Repo
Framework
comments powered by Disqus