July 26, 2019

2115 words 10 mins read

Paper Group NANR 132

Paper Group NANR 132

Unsupervised Aspect Term Extraction with B-LSTM & CRF using Automatically Labelled Datasets. k-Support and Ordered Weighted Sparsity for Overlapping Groups: Hardness and Algorithms. Injecting Word Embeddings with Another Language’s Resource : An Application of Bilingual Embeddings. Lexical Simplification with the Deep Structured Similarity Model. …

Unsupervised Aspect Term Extraction with B-LSTM & CRF using Automatically Labelled Datasets

Title Unsupervised Aspect Term Extraction with B-LSTM & CRF using Automatically Labelled Datasets
Authors Athanasios Giannakopoulos, Claudiu Musat, Andreea Hossmann, Michael Baeriswyl
Abstract Aspect Term Extraction (ATE) identifies opinionated aspect terms in texts and is one of the tasks in the SemEval Aspect Based Sentiment Analysis (ABSA) contest. The small amount of available datasets for supervised ATE and the costly human annotation for aspect term labelling give rise to the need for unsupervised ATE. In this paper, we introduce an architecture that achieves top-ranking performance for supervised ATE. Moreover, it can be used efficiently as feature extractor and classifier for unsupervised ATE. Our second contribution is a method to automatically construct datasets for ATE. We train a classifier on our automatically labelled datasets and evaluate it on the human annotated SemEval ABSA test sets. Compared to a strong rule-based baseline, we obtain a dramatically higher F-score and attain precision values above 80{%}. Our unsupervised method beats the supervised ABSA baseline from SemEval, while preserving high precision scores.
Tasks Aspect-Based Sentiment Analysis, Sentiment Analysis
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-5224/
PDF https://www.aclweb.org/anthology/W17-5224
PWC https://paperswithcode.com/paper/unsupervised-aspect-term-extraction-with-b-1
Repo
Framework

k-Support and Ordered Weighted Sparsity for Overlapping Groups: Hardness and Algorithms

Title k-Support and Ordered Weighted Sparsity for Overlapping Groups: Hardness and Algorithms
Authors Cong Han Lim, Stephen Wright
Abstract The k-support and OWL norms generalize the l1 norm, providing better prediction accuracy and better handling of correlated variables. We study the norms obtained from extending the k-support norm and OWL norms to the setting in which there are overlapping groups. The resulting norms are in general NP-hard to compute, but they are tractable for certain collections of groups. To demonstrate this fact, we develop a dynamic program for the problem of projecting onto the set of vectors supported by a fixed number of groups. Our dynamic program utilizes tree decompositions and its complexity scales with the treewidth. This program can be converted to an extended formulation which, for the associated group structure, models the k-group support norms and an overlapping group variant of the ordered weighted l1 norm. Numerical results demonstrate the efficacy of the new penalties.
Tasks
Published 2017-12-01
URL http://papers.nips.cc/paper/6632-k-support-and-ordered-weighted-sparsity-for-overlapping-groups-hardness-and-algorithms
PDF http://papers.nips.cc/paper/6632-k-support-and-ordered-weighted-sparsity-for-overlapping-groups-hardness-and-algorithms.pdf
PWC https://paperswithcode.com/paper/k-support-and-ordered-weighted-sparsity-for
Repo
Framework

Injecting Word Embeddings with Another Language’s Resource : An Application of Bilingual Embeddings

Title Injecting Word Embeddings with Another Language’s Resource : An Application of Bilingual Embeddings
Authors P, Prakhar ey, Vikram Pudi, Manish Shrivastava
Abstract Word embeddings learned from text corpus can be improved by injecting knowledge from external resources, while at the same time also specializing them for similarity or relatedness. These knowledge resources (like WordNet, Paraphrase Database) may not exist for all languages. In this work we introduce a method to inject word embeddings of a language with knowledge resource of another language by leveraging bilingual embeddings. First we improve word embeddings of German, Italian, French and Spanish using resources of English and test them on variety of word similarity tasks. Then we demonstrate the utility of our method by creating improved embeddings for Urdu and Telugu languages using Hindi WordNet, beating the previously established baseline for Urdu.
Tasks Learning Word Embeddings, Word Embeddings
Published 2017-11-01
URL https://www.aclweb.org/anthology/I17-2020/
PDF https://www.aclweb.org/anthology/I17-2020
PWC https://paperswithcode.com/paper/injecting-word-embeddings-with-another
Repo
Framework

Lexical Simplification with the Deep Structured Similarity Model

Title Lexical Simplification with the Deep Structured Similarity Model
Authors Lis Pereira, Xiaodong Liu, John Lee
Abstract We explore the application of a Deep Structured Similarity Model (DSSM) to ranking in lexical simplification. Our results show that the DSSM can effectively capture fine-grained features to perform semantic matching when ranking substitution candidates, outperforming the state-of-the-art on two standard datasets used for the task.
Tasks Image Captioning, Learning Word Embeddings, Lexical Simplification, Machine Translation, Question Answering, Word Embeddings
Published 2017-11-01
URL https://www.aclweb.org/anthology/I17-2073/
PDF https://www.aclweb.org/anthology/I17-2073
PWC https://paperswithcode.com/paper/lexical-simplification-with-the-deep
Repo
Framework

Joint Concept Learning and Semantic Parsing from Natural Language Explanations

Title Joint Concept Learning and Semantic Parsing from Natural Language Explanations
Authors Shashank Srivastava, Igor Labutov, Tom Mitchell
Abstract Natural language constitutes a predominant medium for much of human learning and pedagogy. We consider the problem of concept learning from natural language explanations, and a small number of labeled examples of the concept. For example, in learning the concept of a phishing email, one might say {`}this is a phishing email because it asks for your bank account number{'}. Solving this problem involves both learning to interpret open ended natural language statements, and learning the concept itself. We present a joint model for (1) language interpretation (semantic parsing) and (2) concept learning (classification) that does not require labeling statements with logical forms. Instead, the model prefers discriminative interpretations of statements in context of observable features of the data as a weak signal for parsing. On a dataset of email-related concepts, our approach yields across-the-board improvements in classification performance, with a 30{%} relative improvement in F1 score over competitive methods in the low data regime. |
Tasks Semantic Parsing
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1161/
PDF https://www.aclweb.org/anthology/D17-1161
PWC https://paperswithcode.com/paper/joint-concept-learning-and-semantic-parsing
Repo
Framework

Beyond Worst-case: A Probabilistic Analysis of Affine Policies in Dynamic Optimization

Title Beyond Worst-case: A Probabilistic Analysis of Affine Policies in Dynamic Optimization
Authors Omar El Housni, Vineet Goyal
Abstract Affine policies (or control) are widely used as a solution approach in dynamic optimization where computing an optimal adjustable solution is usually intractable. While the worst case performance of affine policies can be significantly bad, the empirical performance is observed to be near-optimal for a large class of problem instances. For instance, in the two-stage dynamic robust optimization problem with linear covering constraints and uncertain right hand side, the worst-case approximation bound for affine policies is $O(\sqrt m)$ that is also tight (see Bertsimas and Goyal (2012)), whereas observed empirical performance is near-optimal. In this paper, we aim to address this stark-contrast between the worst-case and the empirical performance of affine policies. In particular, we show that affine policies give a good approximation for the two-stage adjustable robust optimization problem with high probability on random instances where the constraint coefficients are generated i.i.d. from a large class of distributions; thereby, providing a theoretical justification of the observed empirical performance. On the other hand, we also present a distribution such that the performance bound for affine policies on instances generated according to that distribution is $\Omega(\sqrt m)$ with high probability; however, the constraint coefficients are not i.i.d.. This demonstrates that the empirical performance of affine policies can depend on the generative model for instances.
Tasks
Published 2017-12-01
URL http://papers.nips.cc/paper/7061-beyond-worst-case-a-probabilistic-analysis-of-affine-policies-in-dynamic-optimization
PDF http://papers.nips.cc/paper/7061-beyond-worst-case-a-probabilistic-analysis-of-affine-policies-in-dynamic-optimization.pdf
PWC https://paperswithcode.com/paper/beyond-worst-case-a-probabilistic-analysis-of
Repo
Framework

Differentially Private Submodular Maximization: Data Summarization in Disguise

Title Differentially Private Submodular Maximization: Data Summarization in Disguise
Authors Marko Mitrovic, Mark Bun, Andreas Krause, Amin Karbasi
Abstract Many data summarization applications are captured by the general framework of submodular maximization. As a consequence, a wide range of efficient approximation algorithms have been developed. However, when such applications involve sensitive data about individuals, their privacy concerns are not automatically addressed. To remedy this problem, we propose a general and systematic study of differentially private submodular maximization. We present privacy-preserving algorithms for both monotone and non-monotone submodular maximization under cardinality, matroid, and p-extendible system constraints, with guarantees that are competitive with optimal. Along the way, we analyze a new algorithm for non-monotone submodular maximization, which is the first (even non-privately) to achieve a constant approximation ratio while running in linear time. We additionally provide two concrete experiments to validate the efficacy of these algorithms.
Tasks Data Summarization
Published 2017-08-01
URL https://icml.cc/Conferences/2017/Schedule?showEvent=637
PDF http://proceedings.mlr.press/v70/mitrovic17a/mitrovic17a.pdf
PWC https://paperswithcode.com/paper/differentially-private-submodular
Repo
Framework

Convergence rates of a partition based Bayesian multivariate density estimation method

Title Convergence rates of a partition based Bayesian multivariate density estimation method
Authors Linxi Liu, Dangna Li, Wing Hung Wong
Abstract We study a class of non-parametric density estimators under Bayesian settings. The estimators are obtained by adaptively partitioning the sample space. Under a suitable prior, we analyze the concentration rate of the posterior distribution, and demonstrate that the rate does not directly depend on the dimension of the problem in several special cases. Another advantage of this class of Bayesian density estimators is that it can adapt to the unknown smoothness of the true density function, thus achieving the optimal convergence rate without artificial conditions on the density. We also validate the theoretical results on a variety of simulated data sets.
Tasks Density Estimation
Published 2017-12-01
URL http://papers.nips.cc/paper/7059-convergence-rates-of-a-partition-based-bayesian-multivariate-density-estimation-method
PDF http://papers.nips.cc/paper/7059-convergence-rates-of-a-partition-based-bayesian-multivariate-density-estimation-method.pdf
PWC https://paperswithcode.com/paper/convergence-rates-of-a-partition-based
Repo
Framework

Predicting Translation Performance with Referential Translation Machines

Title Predicting Translation Performance with Referential Translation Machines
Authors Ergun Bi{\c{c}}ici
Abstract
Tasks Machine Translation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4759/
PDF https://www.aclweb.org/anthology/W17-4759
PWC https://paperswithcode.com/paper/predicting-translation-performance-with
Repo
Framework

Sheffield MultiMT: Using Object Posterior Predictions for Multimodal Machine Translation

Title Sheffield MultiMT: Using Object Posterior Predictions for Multimodal Machine Translation
Authors Pranava Swaroop Madhyastha, Josiah Wang, Lucia Specia
Abstract
Tasks Image Captioning, Image Classification, Machine Translation, Multimodal Machine Translation, Question Answering, Visual Question Answering
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4752/
PDF https://www.aclweb.org/anthology/W17-4752
PWC https://paperswithcode.com/paper/sheffield-multimt-using-object-posterior
Repo
Framework

NICT-NAIST System for WMT17 Multimodal Translation Task

Title NICT-NAIST System for WMT17 Multimodal Translation Task
Authors Jingyi Zhang, Masao Utiyama, Eiichro Sumita, Graham Neubig, Satoshi Nakamura
Abstract
Tasks Image Retrieval, Machine Translation, Multimodal Machine Translation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4753/
PDF https://www.aclweb.org/anthology/W17-4753
PWC https://paperswithcode.com/paper/nict-naist-system-for-wmt17-multimodal
Repo
Framework

Interactive Submodular Bandit

Title Interactive Submodular Bandit
Authors Lin Chen, Andreas Krause, Amin Karbasi
Abstract In many machine learning applications, submodular functions have been used as a model for evaluating the utility or payoff of a set such as news items to recommend, sensors to deploy in a terrain, nodes to influence in a social network, to name a few. At the heart of all these applications is the assumption that the underlying utility/payoff function is known a priori, hence maximizing it is in principle possible. In real life situations, however, the utility function is not fully known in advance and can only be estimated via interactions. For instance, whether a user likes a movie or not can be reliably evaluated only after it was shown to her. Or, the range of influence of a user in a social network can be estimated only after she is selected to advertise the product. We model such problems as an interactive submodular bandit optimization, where in each round we receive a context (e.g., previously selected movies) and have to choose an action (e.g., propose a new movie). We then receive a noisy feedback about the utility of the action (e.g., ratings) which we model as a submodular function over the context-action space. We develop SM-UCB that efficiently trades off exploration (collecting more data) and exploration (proposing a good action given gathered data) and achieves a $O(\sqrt{T})$ regret bound after $T$ rounds of interaction. Given a bounded-RKHS norm kernel over the context-action-payoff space that governs the smoothness of the utility function, SM-UCB keeps an upper-confidence bound on the payoff function that allows it to asymptotically achieve no-regret. Finally, we evaluate our results on four concrete applications, including movie recommendation (on the MovieLense data set), news recommendation (on Yahoo! Webscope dataset), interactive influence maximization (on a subset of the Facebook network), and personalized data summarization (on Reuters Corpus). In all these applications, we observe that SM-UCB consistently outperforms the prior art.
Tasks Data Summarization
Published 2017-12-01
URL http://papers.nips.cc/paper/6619-interactive-submodular-bandit
PDF http://papers.nips.cc/paper/6619-interactive-submodular-bandit.pdf
PWC https://paperswithcode.com/paper/interactive-submodular-bandit
Repo
Framework

Methodical Evaluation of Arabic Word Embeddings

Title Methodical Evaluation of Arabic Word Embeddings
Authors Mohammed Elrazzaz, Shady Elbassuoni, Khaled Shaban, Chadi Helwe
Abstract Many unsupervised learning techniques have been proposed to obtain meaningful representations of words from text. In this study, we evaluate these various techniques when used to generate Arabic word embeddings. We first build a benchmark for the Arabic language that can be utilized to perform intrinsic evaluation of different word embeddings. We then perform additional extrinsic evaluations of the embeddings based on two NLP tasks.
Tasks Document Classification, Learning Word Embeddings, Named Entity Recognition, Word Embeddings
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-2072/
PDF https://www.aclweb.org/anthology/P17-2072
PWC https://paperswithcode.com/paper/methodical-evaluation-of-arabic-word
Repo
Framework

Bilexical Embeddings for Quality Estimation

Title Bilexical Embeddings for Quality Estimation
Authors Fr{'e}d{'e}ric Blain, Carolina Scarton, Lucia Specia
Abstract
Tasks Language Modelling, Machine Translation, Word Embeddings
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4760/
PDF https://www.aclweb.org/anthology/W17-4760
PWC https://paperswithcode.com/paper/bilexical-embeddings-for-quality-estimation
Repo
Framework

Improving Machine Translation Quality Estimation with Neural Network Features

Title Improving Machine Translation Quality Estimation with Neural Network Features
Authors Zhiming Chen, Yiming Tan, Chenlin Zhang, Qingyu Xiang, Lilin Zhang, Maoxi Li, Mingwen Wang
Abstract
Tasks Language Modelling, Machine Translation, Sentence Embedding, Speech Recognition
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4761/
PDF https://www.aclweb.org/anthology/W17-4761
PWC https://paperswithcode.com/paper/improving-machine-translation-quality
Repo
Framework
comments powered by Disqus