July 26, 2019

2564 words 13 mins read

Paper Group NANR 103

Paper Group NANR 103

The AFRL WMT17 Neural Machine Translation Training Task Submission. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Efficient Nonmyopic Active Search. How much progress have we made on RST discourse parsing? A replication study of recent results on the RST-DT. Do Not Trust the Trolls: …

The AFRL WMT17 Neural Machine Translation Training Task Submission

Title The AFRL WMT17 Neural Machine Translation Training Task Submission
Authors Jeremy Gwinnup, Grant Erdmann, Katherine Young
Abstract
Tasks Machine Translation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4781/
PDF https://www.aclweb.org/anthology/W17-4781
PWC https://paperswithcode.com/paper/the-afrl-wmt17-neural-machine-translation
Repo
Framework

Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)

Title Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Authors
Abstract
Tasks
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-1000/
PDF https://www.aclweb.org/anthology/P17-1000
PWC https://paperswithcode.com/paper/proceedings-of-the-55th-annual-meeting-of-the-1
Repo
Framework
Title Efficient Nonmyopic Active Search
Authors Shali Jiang, Gustavo Malkomes, Geoff Converse, Alyssa Shofner, Benjamin Moseley, Roman Garnett
Abstract Active search is an active learning setting with the goal of identifying as many members of a given class as possible under a labeling budget. In this work, we first establish a theoretical hardness of active search, proving that no polynomial-time policy can achieve a constant factor approximation ratio with respect to the expected utility of the optimal policy. We also propose a novel, computationally efficient active search policy achieving exceptional performance on several real-world tasks. Our policy is nonmyopic, always considering the entire remaining search budget. It also automatically and dynamically balances exploration and exploitation consistent with the remaining budget, without relying on a parameter to control this tradeoff. We conduct experiments on diverse datasets from several domains: drug discovery, materials science, and a citation network. Our efficient nonmyopic policy recovers significantly more valuable points with the same budget than several alternatives from the literature, including myopic approximations to the optimal policy.
Tasks Active Learning, Drug Discovery
Published 2017-08-01
URL https://icml.cc/Conferences/2017/Schedule?showEvent=553
PDF http://proceedings.mlr.press/v70/jiang17d/jiang17d.pdf
PWC https://paperswithcode.com/paper/efficient-nonmyopic-active-search
Repo
Framework

How much progress have we made on RST discourse parsing? A replication study of recent results on the RST-DT

Title How much progress have we made on RST discourse parsing? A replication study of recent results on the RST-DT
Authors Mathieu Morey, Philippe Muller, Nicholas Asher
Abstract This article evaluates purported progress over the past years in RST discourse parsing. Several studies report a relative error reduction of 24 to 51{%} on all metrics that authors attribute to the introduction of distributed representations of discourse units. We replicate the standard evaluation of 9 parsers, 5 of which use distributed representations, from 8 studies published between 2013 and 2017, using their predictions on the test set of the RST-DT. Our main finding is that most recently reported increases in RST discourse parser performance are an artefact of differences in implementations of the evaluation procedure. We evaluate all these parsers with the standard Parseval procedure to provide a more accurate picture of the actual RST discourse parsers performance in standard evaluation settings. Under this more stringent procedure, the gains attributable to distributed representations represent at most a 16{%} relative error reduction on fully-labelled structures.
Tasks
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1136/
PDF https://www.aclweb.org/anthology/D17-1136
PWC https://paperswithcode.com/paper/how-much-progress-have-we-made-on-rst
Repo
Framework

Do Not Trust the Trolls: Predicting Credibility in Community Question Answering Forums

Title Do Not Trust the Trolls: Predicting Credibility in Community Question Answering Forums
Authors Preslav Nakov, Tsvetomila Mihaylova, Llu{'\i}s M{`a}rquez, Yashkumar Shiroya, Ivan Koychev
Abstract We address information credibility in community forums, in a setting in which the credibility of an answer posted in a question thread by a particular user has to be predicted. First, we motivate the problem and we create a publicly available annotated English corpus by crowdsourcing. Second, we propose a large set of features to predict the credibility of the answers. The features model the user, the answer, the question, the thread as a whole, and the interaction between them. Our experiments with ranking SVMs show that the credibility labels can be predicted with high performance according to several standard IR ranking metrics, thus supporting the potential usage of this layer of credibility information in practical applications. The features modeling the profile of the user (in particular trollness) turn out to be most important, but embedding features modeling the answer and the similarity between the question and the answer are also very relevant. Overall, half of the gap between the baseline performance and the perfect classifier can be covered using the proposed features.
Tasks Community Question Answering, Information Retrieval, Question Answering
Published 2017-09-01
URL https://www.aclweb.org/anthology/R17-1072/
PDF https://doi.org/10.26615/978-954-452-049-6_072
PWC https://paperswithcode.com/paper/do-not-trust-the-trolls-predicting
Repo
Framework

Elementary Symmetric Polynomials for Optimal Experimental Design

Title Elementary Symmetric Polynomials for Optimal Experimental Design
Authors Zelda E. Mariet, Suvrit Sra
Abstract We revisit the classical problem of optimal experimental design (OED) under a new mathematical model grounded in a geometric motivation. Specifically, we introduce models based on elementary symmetric polynomials; these polynomials capture “partial volumes” and offer a graded interpolation between the widely used A-optimal and D-optimal design models, obtaining each of them as special cases. We analyze properties of our models, and derive both greedy and convex-relaxation algorithms for computing the associated designs. Our analysis establishes approximation guarantees on these algorithms, while our empirical results substantiate our claims and demonstrate a curious phenomenon concerning our greedy algorithm. Finally, as a byproduct, we obtain new results on the theory of elementary symmetric polynomials that may be of independent interest.
Tasks
Published 2017-12-01
URL http://papers.nips.cc/paper/6809-elementary-symmetric-polynomials-for-optimal-experimental-design
PDF http://papers.nips.cc/paper/6809-elementary-symmetric-polynomials-for-optimal-experimental-design.pdf
PWC https://paperswithcode.com/paper/elementary-symmetric-polynomials-for-optimal
Repo
Framework

IIT-UHH at SemEval-2017 Task 3: Exploring Multiple Features for Community Question Answering and Implicit Dialogue Identification

Title IIT-UHH at SemEval-2017 Task 3: Exploring Multiple Features for Community Question Answering and Implicit Dialogue Identification
Authors N, Titas i, Chris Biemann, Seid Muhie Yimam, Deepak Gupta, Sarah Kohail, Asif Ekbal, Pushpak Bhattacharyya
Abstract In this paper we present the system for Answer Selection and Ranking in Community Question Answering, which we build as part of our participation in SemEval-2017 Task 3. We develop a Support Vector Machine (SVM) based system that makes use of textual, domain-specific, word-embedding and topic-modeling features. In addition, we propose a novel method for dialogue chain identification in comment threads. Our primary submission won subtask C, outperforming other systems in all the primary evaluation metrics. We performed well in other English subtasks, ranking third in subtask A and eighth in subtask B. We also developed open source toolkits for all the three English subtasks by the name cQARank [\url{https://github.com/TitasNandi/cQARank}].
Tasks Answer Selection, Community Question Answering, Question Answering, Question Similarity
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2009/
PDF https://www.aclweb.org/anthology/S17-2009
PWC https://paperswithcode.com/paper/iit-uhh-at-semeval-2017-task-3-exploring
Repo
Framework

SCIR-QA at SemEval-2017 Task 3: CNN Model Based on Similar and Dissimilar Information between Keywords for Question Similarity

Title SCIR-QA at SemEval-2017 Task 3: CNN Model Based on Similar and Dissimilar Information between Keywords for Question Similarity
Authors Le Qi, Yu Zhang, Ting Liu
Abstract We describe a method of calculating the similarity of questions in community QA. Question in cQA are usually very long and there are a lot of useless information about calculating the similarity of questions. Therefore,we implement a CNN model based on similar and dissimilar information between question{'}s keywords. We extract the keywords of questions, and then model the similar and dissimilar information between the keywords, and use the CNN model to calculate the similarity.
Tasks Community Question Answering, Graph Ranking, Information Retrieval, Keyword Extraction, Question Answering, Question Similarity
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2049/
PDF https://www.aclweb.org/anthology/S17-2049
PWC https://paperswithcode.com/paper/scir-qa-at-semeval-2017-task-3-cnn-model
Repo
Framework

SimBow at SemEval-2017 Task 3: Soft-Cosine Semantic Similarity between Questions for Community Question Answering

Title SimBow at SemEval-2017 Task 3: Soft-Cosine Semantic Similarity between Questions for Community Question Answering
Authors Delphine Charlet, G{'e}raldine Damnati
Abstract This paper describes the SimBow system submitted at SemEval2017-Task3, for the question-question similarity subtask B. The proposed approach is a supervised combination of different unsupervised textual similarities. These textual similarities rely on the introduction of a relation matrix in the classical cosine similarity between bag-of-words, so as to get a soft-cosine that takes into account relations between words. According to the type of relation matrix embedded in the soft-cosine, semantic or lexical relations can be considered. Our system ranked first among the official submissions of subtask B.
Tasks Community Question Answering, Knowledge Graphs, Question Answering, Question Similarity, Semantic Similarity, Semantic Textual Similarity, Word Embeddings
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2051/
PDF https://www.aclweb.org/anthology/S17-2051
PWC https://paperswithcode.com/paper/simbow-at-semeval-2017-task-3-soft-cosine
Repo
Framework

NLM_NIH at SemEval-2017 Task 3: from Question Entailment to Question Similarity for Community Question Answering

Title NLM_NIH at SemEval-2017 Task 3: from Question Entailment to Question Similarity for Community Question Answering
Authors Asma Ben Abacha, Dina Demner-Fushman
Abstract This paper describes our participation in SemEval-2017 Task 3 on Community Question Answering (cQA). The Question Similarity subtask (B) aims to rank a set of related questions retrieved by a search engine according to their similarity to the original question. We adapted our feature-based system for Recognizing Question Entailment (RQE) to the question similarity task. Tested on cQA-B-2016 test data, our RQE system outperformed the best system of the 2016 challenge in all measures with 77.47 MAP and 80.57 Accuracy. On cQA-B-2017 test data, performances of all systems dropped by around 30 points. Our primary system obtained 44.62 MAP, 67.27 Accuracy and 47.25 F1 score. The cQA-B-2017 best system achieved 47.22 MAP and 42.37 F1 score. Our system is ranked sixth in terms of MAP and third in terms of F1 out of 13 participating teams.
Tasks Community Question Answering, Question Answering, Question Similarity
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2057/
PDF https://www.aclweb.org/anthology/S17-2057
PWC https://paperswithcode.com/paper/nlm_nih-at-semeval-2017-task-3-from-question
Repo
Framework

Source-Target Similarity Modelings for Multi-Source Transfer Gaussian Process Regression

Title Source-Target Similarity Modelings for Multi-Source Transfer Gaussian Process Regression
Authors Pengfei Wei, Ramon Sagarna, Yiping Ke, Yew-Soon Ong, Chi-Keong Goh
Abstract A key challenge in multi-source transfer learning is to capture the diverse inter-domain similarities. In this paper, we study different approaches based on Gaussian process models to solve the multi-source transfer regression problem. Precisely, we first investigate the feasibility and performance of a family of transfer covariance functions that represent the pairwise similarity of each source and the target domain. We theoretically show that using such a transfer covariance function for general Gaussian process modelling can only capture the same similarity coefficient for all the sources, and thus may result in unsatisfactory transfer performance. This leads us to propose TC$_{MS}$Stack, an integrated strategy incorporating the benefits of the transfer covariance function and stacking. Extensive experiments on one synthetic and two real-world datasets, with learning settings of up to 11 sources for the latter, demonstrate the effectiveness of our proposed TC$_{MS}$Stack.
Tasks Transfer Learning
Published 2017-08-01
URL https://icml.cc/Conferences/2017/Schedule?showEvent=679
PDF http://proceedings.mlr.press/v70/wei17a/wei17a.pdf
PWC https://paperswithcode.com/paper/source-target-similarity-modelings-for-multi
Repo
Framework

ECNU at SemEval-2017 Task 3: Using Traditional and Deep Learning Methods to Address Community Question Answering Task

Title ECNU at SemEval-2017 Task 3: Using Traditional and Deep Learning Methods to Address Community Question Answering Task
Authors Guoshun Wu, Yixuan Sheng, Man Lan, Yuanbin Wu
Abstract This paper describes the systems we submitted to the task 3 (Community Question Answering) in SemEval 2017 which contains three subtasks on English corpora, i.e., subtask A: Question-Comment Similarity, subtask B: Question-Question Similarity, and subtask C: Question-External Comment Similarity. For subtask A, we combined two different methods to represent question-comment pair, i.e., supervised model using traditional features and Convolutional Neural Network. For subtask B, we utilized the information of snippets returned from Search Engine with question subject as query. For subtask C, we ranked the comments by multiplying the probability of the pair related question comment being Good by the reciprocal rank of the related question.
Tasks Community Question Answering, Question Answering, Question Similarity, Semantic Textual Similarity
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2060/
PDF https://www.aclweb.org/anthology/S17-2060
PWC https://paperswithcode.com/paper/ecnu-at-semeval-2017-task-3-using-traditional
Repo
Framework

Probabilistic Temporal Subspace Clustering

Title Probabilistic Temporal Subspace Clustering
Authors Behnam Gholami, Vladimir Pavlovic
Abstract Subspace clustering is a common modeling paradigm used to identify constituent modes of variation in data with locally linear structure. These structures are common to many problems in computer vision, including modeling time series of complex human motion. However classical subspace clustering algorithms learn the relationships within a set of data without considering the temporal dependency and then use a separate clustering step (e.g., spectral clustering) for final segmentation. Moreover, these, frequently optimization-based, algorithms assume that all observations have complete features. In contrast in real-world applications, some features are often missing, which results in incomplete data and substantial performance degeneration of these approaches. In this paper, we propose a unified non-parametric generative framework for temporal subspace clustering to segment data drawn from a sequentially ordered union of subspaces that deals with the missing features in a principled way. The non-parametric nature of our generative model makes it possible to infer the number of subspaces and their dimension automatically from data. Experimental results on human action datasets demonstrate that the proposed model consistently outperforms other state-of-the-art subspace clustering approaches.
Tasks Time Series
Published 2017-07-01
URL http://openaccess.thecvf.com/content_cvpr_2017/html/Gholami_Probabilistic_Temporal_Subspace_CVPR_2017_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2017/papers/Gholami_Probabilistic_Temporal_Subspace_CVPR_2017_paper.pdf
PWC https://paperswithcode.com/paper/probabilistic-temporal-subspace-clustering
Repo
Framework

Semantic Vector Encoding and Similarity Search Using Fulltext Search Engines

Title Semantic Vector Encoding and Similarity Search Using Fulltext Search Engines
Authors Jan Rygl, Jan Pomik{'a}lek, Radim {\v{R}}eh{\r{u}}{\v{r}}ek, Michal R{\r{u}}{\v{z}}i{\v{c}}ka, V{'\i}t Novotn{'y}, Petr Sojka
Abstract Vector representations and vector space modeling (VSM) play a central role in modern machine learning. We propose a novel approach to {`}vector similarity searching{'} over dense semantic representations of words and documents that can be deployed on top of traditional inverted-index-based fulltext engines, taking advantage of their robustness, stability, scalability and ubiquity. We show that this approach allows the indexing and querying of dense vectors in text domains. This opens up exciting avenues for major efficiency gains, along with simpler deployment, scaling and monitoring. The end result is a fast and scalable vector database with a tunable trade-off between vector search performance and quality, backed by a standard fulltext engine such as Elasticsearch. We empirically demonstrate its querying performance and quality by applying this solution to the task of semantic searching over a dense vector representation of the entire English Wikipedia. |
Tasks Information Retrieval, Representation Learning
Published 2017-08-01
URL https://www.aclweb.org/anthology/W17-2611/
PDF https://www.aclweb.org/anthology/W17-2611
PWC https://paperswithcode.com/paper/semantic-vector-encoding-and-similarity
Repo
Framework

Incorporating Uncertainty into Deep Learning for Spoken Language Assessment

Title Incorporating Uncertainty into Deep Learning for Spoken Language Assessment
Authors Andrey Malinin, Anton Ragni, Kate Knill, Mark Gales
Abstract There is a growing demand for automatic assessment of spoken English proficiency. These systems need to handle large variations in input data owing to the wide range of candidate skill levels and L1s, and errors from ASR. Some candidates will be a poor match to the training data set, undermining the validity of the predicted grade. For high stakes tests it is essential for such systems not only to grade well, but also to provide a measure of their uncertainty in their predictions, enabling rejection to human graders. Previous work examined Gaussian Process (GP) graders which, though successful, do not scale well with large data sets. Deep Neural Network (DNN) may also be used to provide uncertainty using Monte-Carlo Dropout (MCD). This paper proposes a novel method to yield uncertainty and compares it to GPs and DNNs with MCD. The proposed approach explicitly teaches a DNN to have low uncertainty on training data and high uncertainty on generated artificial data. On experiments conducted on data from the Business Language Testing Service (BULATS), the proposed approach is found to outperform GPs and DNNs with MCD in uncertainty-based rejection whilst achieving comparable grading performance.
Tasks
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-2008/
PDF https://www.aclweb.org/anthology/P17-2008
PWC https://paperswithcode.com/paper/incorporating-uncertainty-into-deep-learning
Repo
Framework
comments powered by Disqus