January 24, 2020

2405 words 12 mins read

Paper Group NANR 197

Paper Group NANR 197

WordNet Gloss Translation for Under-resourced Languages using Multilingual Neural Machine Translation. Would MT kill creativity in literary retranslation?. Building a Speech Corpus based on Arabic Podcasts for Language and Dialect Identification. AspeRa: Aspect-Based Rating Prediction Based on User Reviews. Drive&Act: A Multi-Modal Dataset for Fine …

WordNet Gloss Translation for Under-resourced Languages using Multilingual Neural Machine Translation

Title WordNet Gloss Translation for Under-resourced Languages using Multilingual Neural Machine Translation
Authors Bharathi Raja Chakravarthi, Mihael Arcan, John P. McCrae
Abstract
Tasks Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-7101/
PDF https://www.aclweb.org/anthology/W19-7101
PWC https://paperswithcode.com/paper/wordnet-gloss-translation-for-under-resourced
Repo
Framework

Would MT kill creativity in literary retranslation?

Title Would MT kill creativity in literary retranslation?
Authors Mehmet {\c{S}}ahin, Sabri G{"u}rses
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-7304/
PDF https://www.aclweb.org/anthology/W19-7304
PWC https://paperswithcode.com/paper/would-mt-kill-creativity-in-literary
Repo
Framework

Building a Speech Corpus based on Arabic Podcasts for Language and Dialect Identification

Title Building a Speech Corpus based on Arabic Podcasts for Language and Dialect Identification
Authors Khaled Lounnas, Mourad Abbas, Mohamed Lichouri
Abstract
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-7408/
PDF https://www.aclweb.org/anthology/W19-7408
PWC https://paperswithcode.com/paper/building-a-speech-corpus-based-on-arabic
Repo
Framework

AspeRa: Aspect-Based Rating Prediction Based on User Reviews

Title AspeRa: Aspect-Based Rating Prediction Based on User Reviews
Authors Elena Tutubalina, Valentin Malykh, Sergey Nikolenko, Anton Alekseev, Ilya Shenbin
Abstract We propose a novel Aspect-based Rating Prediction model (AspeRa) that estimates user rating based on review texts for the items. It is based on aspect extraction with neural networks and combines the advantages of deep learning and topic modeling. It is mainly designed for recommendations, but an important secondary goal of AspeRa is to discover coherent aspects of reviews that can be used to explain predictions or for user profiling. We conduct a comprehensive empirical study of AspeRa, showing that it outperforms state-of-the-art models in terms of recommendation quality and produces interpretable aspects. This paper is an abridged version of our work (Nikolenko et al., 2019)
Tasks Aspect Extraction
Published 2019-08-01
URL https://www.aclweb.org/anthology/papers/W/W19/W19-3605/
PDF https://www.aclweb.org/anthology/W19-3605
PWC https://paperswithcode.com/paper/aspera-aspect-based-rating-prediction-based
Repo
Framework

Drive&Act: A Multi-Modal Dataset for Fine-Grained Driver Behavior Recognition in Autonomous Vehicles

Title Drive&Act: A Multi-Modal Dataset for Fine-Grained Driver Behavior Recognition in Autonomous Vehicles
Authors Manuel Martin, Alina Roitberg, Monica Haurilet, Matthias Horne, Simon Reiss, Michael Voit, Rainer Stiefelhagen
Abstract We introduce the novel domain-specific Drive&Act benchmark for fine-grained categorization of driver behavior. Our dataset features twelve hours and over 9.6 million frames of people engaged in distractive activities during both, manual and automated driving. We capture color, infrared, depth and 3D body pose information from six views and densely label the videos with a hierarchical annotation scheme, resulting in 83 categories. The key challenges of our dataset are: (1) recognition of fine-grained behavior inside the vehicle cabin; (2) multi-modal activity recognition, focusing on diverse data streams; and (3) a cross view recognition benchmark, where a model handles data from an unfamiliar domain, as sensor type and placement in the cabin can change between vehicles. Finally, we provide challenging benchmarks by adopting prominent methods for video- and body pose-based action recognition.
Tasks Activity Recognition, Autonomous Vehicles
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Martin_DriveAct_A_Multi-Modal_Dataset_for_Fine-Grained_Driver_Behavior_Recognition_in_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Martin_DriveAct_A_Multi-Modal_Dataset_for_Fine-Grained_Driver_Behavior_Recognition_in_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/driveact-a-multi-modal-dataset-for-fine
Repo
Framework

Fermi at SemEval-2019 Task 8: An elementary but effective approach to Question Discernment in Community QA Forums

Title Fermi at SemEval-2019 Task 8: An elementary but effective approach to Question Discernment in Community QA Forums
Authors Bakhtiyar Syed, Vijayasaradhi Indurthi, Manish Shrivastava, Manish Gupta, Vasudeva Varma
Abstract Online Community Question Answering Forums (cQA) have gained massive popularity within recent years. The rise in users for such forums have led to the increase in the need for automated evaluation for question comprehension and fact evaluation of the answers provided by various participants in the forum. Our team, \textbf{Fermi}, participated in sub-task A of Task 8 at SemEval 2019 - which tackles the first problem in the pipeline of factual evaluation in cQA forums, i.e., deciding whether a posed question asks for a factual information, an opinion/advice or is just socializing. This information is highly useful in segregating factual questions from non-factual ones which highly helps in organizing the questions into useful categories and trims down the problem space for the next task in the pipeline for fact evaluation among the available answers. Our system uses the embeddings obtained from Universal Sentence Encoder combined with XGBoost for the classification sub-task A. We also evaluate other combinations of embeddings and off-the-shelf machine learning algorithms to demonstrate the efficacy of the various representations and their combinations. Our results across the evaluation test set gave an accuracy of 84{%} and received the first position in the final standings judged by the organizers.
Tasks Community Question Answering, Question Answering
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2203/
PDF https://www.aclweb.org/anthology/S19-2203
PWC https://paperswithcode.com/paper/fermi-at-semeval-2019-task-8-an-elementary
Repo
Framework

MVP Matching: A Maximum-Value Perfect Matching for Mining Hard Samples, With Application to Person Re-Identification

Title MVP Matching: A Maximum-Value Perfect Matching for Mining Hard Samples, With Application to Person Re-Identification
Authors Han Sun, Zhiyuan Chen, Shiyang Yan, Lin Xu
Abstract How to correctly stress hard samples in metric learning is critical for visual recognition tasks, especially in challenging person re-ID applications. Pedestrians across cameras with significant appearance variations are easily confused, which could bias the learned metric and slow down the convergence rate. In this paper, we propose a novel weighted complete bipartite graph based maximum-value perfect (MVP) matching for mining the hard samples from a batch of samples. It can emphasize the hard positive and negative sample pairs respectively, and thus relieve adverse optimization and sample imbalance problems. We then develop a new batch-wise MVP matching based loss objective and combine it in an end-to-end deep metric learning manner. It leads to significant improvements in both convergence rate and recognition performance. Extensive empirical results on five person re-ID benchmark datasets, i.e., Market-1501, CUHK03-Detected, CUHK03-Labeled, Duke-MTMC, and MSMT17, demonstrate the superiority of the proposed method. It can accelerate the convergence rate significantly while achieving state-of-the-art performance. The source code of our method is available at https://github.com/IAAI-CVResearchGroup/MVP-metric.
Tasks Metric Learning, Person Re-Identification
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Sun_MVP_Matching_A_Maximum-Value_Perfect_Matching_for_Mining_Hard_Samples_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Sun_MVP_Matching_A_Maximum-Value_Perfect_Matching_for_Mining_Hard_Samples_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/mvp-matching-a-maximum-value-perfect-matching
Repo
Framework

Cross-Sentence Transformations in Text Simplification

Title Cross-Sentence Transformations in Text Simplification
Authors Fern Alva-Manchego, o, Carolina Scarton, Lucia Specia
Abstract Current approaches to Text Simplification focus on simplifying sentences individually. However, certain simplification transformations span beyond single sentences (e.g. joining and re- ordering sentences). In this paper, we motivate the need for modelling the simplification task at the document level, and assess the performance of sequence-to-sequence neural models in this setup. We analyse parallel original-simplified documents created by professional editors and show that there are frequent rewriting transformations that are not restricted to sentence boundaries. We also propose strategies to automatically evaluate the performance of a simplification model on these cross-sentence transformations. Our experiments show the inability of standard sequence-to-sequence neural models to learn these transformations, and suggest directions towards document-level simplification.
Tasks Text Simplification
Published 2019-08-01
URL https://www.aclweb.org/anthology/papers/W/W19/W19-3656/
PDF https://www.aclweb.org/anthology/W19-3656
PWC https://paperswithcode.com/paper/cross-sentence-transformations-in-text
Repo
Framework

DeepGeneMD: A Joint Deep Learning Model for Extracting Gene Mutation-Disease Knowledge from PubMed Literature

Title DeepGeneMD: A Joint Deep Learning Model for Extracting Gene Mutation-Disease Knowledge from PubMed Literature
Authors Feifan Liu, Xiaoyu Zheng, Bo Wang, Catarina Kiefe
Abstract Understanding the pathogenesis of genetic diseases through different gene activities and their relations to relevant diseases is important for new drug discovery and drug repositioning. In this paper, we present a joint deep learning model in a multi-task learning paradigm for gene mutation-disease knowledge extraction, DeepGeneMD, which adapts the state-of-the-art hierarchical multi-task learning framework for joint inference on named entity recognition (NER) and relation extraction (RE) in the context of the AGAC (Active Gene Annotation Corpus) track at 2019 BioNLP Open Shared Tasks (BioNLP-OST). It simultaneously extracts gene mutation related activities, diseases, and their relations from the published scientific literature. In DeepGeneMD, we explore the task decomposition to create auxiliary subtasks so that more interactions between different learning subtasks can be leveraged in model training. Our model achieves the average F1 score of 0.45 on recognizing gene activities and disease entities, ranking 2nd in the AGAC NER task; and the average F1 score of 0.35 on extracting relations, ranking 1st in the AGAC RE task.
Tasks Drug Discovery, Multi-Task Learning, Named Entity Recognition, Relation Extraction
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5712/
PDF https://www.aclweb.org/anthology/D19-5712
PWC https://paperswithcode.com/paper/deepgenemd-a-joint-deep-learning-model-for
Repo
Framework

Not All Reviews Are Equal: Towards Addressing Reviewer Biases for Opinion Summarization

Title Not All Reviews Are Equal: Towards Addressing Reviewer Biases for Opinion Summarization
Authors Wenyi Tay
Abstract Consumers read online reviews for insights which help them to make decisions. Given the large volumes of reviews, succinct review summaries are important for many applications. Existing research has focused on mining for opinions from only review texts and largely ignores the reviewers. However, reviewers have biases and may write lenient or harsh reviews; they may also have preferences towards some topics over others. Therefore, not all reviews are equal. Ignoring the biases in reviews can generate misleading summaries. We aim for summarization of reviews to include balanced opinions from reviewers of different biases and preferences. We propose to model reviewer biases from their review texts and rating distributions, and learn a bias-aware opinion representation. We further devise an approach for balanced opinion summarization of reviews using our bias-aware opinion representation.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-2005/
PDF https://www.aclweb.org/anthology/P19-2005
PWC https://paperswithcode.com/paper/not-all-reviews-are-equal-towards-addressing
Repo
Framework

On a Chatbot Conducting Dialogue-in-Dialogue

Title On a Chatbot Conducting Dialogue-in-Dialogue
Authors Boris Galitsky, Dmitry Ilvovsky, Elizaveta Goncharova
Abstract We demo a chatbot that delivers content in the form of virtual dialogues automatically produced from plain texts extracted and selected from documents. This virtual dialogue content is provided in the form of answers derived from the found and selected documents split into fragments, and questions are automatically generated for these answers.
Tasks Chatbot
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-5916/
PDF https://www.aclweb.org/anthology/W19-5916
PWC https://paperswithcode.com/paper/on-a-chatbot-conducting-dialogue-in-dialogue
Repo
Framework

Transformer based Grapheme-to-Phoneme Conversion

Title Transformer based Grapheme-to-Phoneme Conversion
Authors Sevinj Yolchuyeva, Géza Németh, Bálint Gyires-Tóth
Abstract Attention mechanism is one of the most successful techniques in deep learning based Natural Language Processing (NLP). The transformer network architecture is completely based on attention mechanisms, and it outperforms sequence-to-sequence models in neural machine translation without recurrent and convolutional layers. Grapheme-to-phoneme (G2P) conversion is a task of converting letters (grapheme sequence) to their pronunciations (phoneme sequence). It plays a significant role in text-to-speech (TTS) and automatic speech recognition (ASR) systems. In this paper, we investigate the application of transformer architecture to G2P conversion and compare its performance with recurrent and convolutional neural network based approaches. Phoneme and word error rates are evaluated on the CMUDict dataset for US English and the NetTalk dataset. The results show that transformer based G2P outperforms the convolutional-based approach in terms of word error rate and our results significantly exceeded previous recurrent approaches (without attention) regarding word and phoneme error rates on both datasets. Furthermore, the size of the proposed model is much smaller than the size of the previous approaches.
Tasks Machine Translation, Speech Recognition
Published 2019-09-15
URL https://pdfs.semanticscholar.org/5bdd/df7eac1d872d2d720be0d28f22cf65c22d2b.pdf
PDF https://pdfs.semanticscholar.org/5bdd/df7eac1d872d2d720be0d28f22cf65c22d2b.pdf
PWC https://paperswithcode.com/paper/transformer-based-grapheme-to-phoneme
Repo
Framework

Meta-Learning to Guide Segmentation

Title Meta-Learning to Guide Segmentation
Authors Kate Rakelly*, Evan Shelhamer*, Trevor Darrell, Alexei A. Efros, Sergey Levine
Abstract There are myriad kinds of segmentation, and ultimately the `"right” segmentation of a given scene is in the eye of the annotator. Standard approaches require large amounts of labeled data to learn just one particular kind of segmentation. As a first step towards relieving this annotation burden, we propose the problem of guided segmentation: given varying amounts of pixel-wise labels, segment unannotated pixels by propagating supervision locally (within an image) and non-locally (across images). We propose guided networks, which extract a latent task representation—guidance—from variable amounts and classes (categories, instances, etc.) of pixel supervision and optimize our architecture end-to-end for fast, accurate, and data-efficient segmentation by meta-learning. To span the few-shot and many-shot learning regimes, we examine guidance from as little as one pixel per concept to as much as 1000+ images, and compare to full gradient optimization at both extremes. To explore generalization, we analyze guidance as a bridge between different levels of supervision to segment classes as the union of instances. Our segmentor concentrates different amounts of supervision of different types of classes into an efficient latent representation, non-locally propagates this supervision across images, and can be updated quickly and cumulatively when given more supervision. |
Tasks Meta-Learning
Published 2019-05-01
URL https://openreview.net/forum?id=HJej6jR5Fm
PDF https://openreview.net/pdf?id=HJej6jR5Fm
PWC https://paperswithcode.com/paper/meta-learning-to-guide-segmentation
Repo
Framework

Multi-Modal Sarcasm Detection in Twitter with Hierarchical Fusion Model

Title Multi-Modal Sarcasm Detection in Twitter with Hierarchical Fusion Model
Authors Yitao Cai, Huiyu Cai, Xiaojun Wan
Abstract Sarcasm is a subtle form of language in which people express the opposite of what is implied. Previous works of sarcasm detection focused on texts. However, more and more social media platforms like Twitter allow users to create multi-modal messages, including texts, images, and videos. It is insufficient to detect sarcasm from multi-model messages based only on texts. In this paper, we focus on multi-modal sarcasm detection for tweets consisting of texts and images in Twitter. We treat text features, image features and image attributes as three modalities and propose a multi-modal hierarchical fusion model to address this task. Our model first extracts image features and attribute features, and then leverages attribute features and bidirectional LSTM network to extract text features. Features of three modalities are then reconstructed and fused into one feature vector for prediction. We create a multi-modal sarcasm detection dataset based on Twitter. Evaluation results on the dataset demonstrate the efficacy of our proposed model and the usefulness of the three modalities.
Tasks Sarcasm Detection
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1239/
PDF https://www.aclweb.org/anthology/P19-1239
PWC https://paperswithcode.com/paper/multi-modal-sarcasm-detection-in-twitter-with
Repo
Framework

Adversarial Attention Modeling for Multi-dimensional Emotion Regression

Title Adversarial Attention Modeling for Multi-dimensional Emotion Regression
Authors Suyang Zhu, Shoushan Li, Guodong Zhou
Abstract In this paper, we propose a neural network-based approach, namely Adversarial Attention Network, to the task of multi-dimensional emotion regression, which automatically rates multiple emotion dimension scores for an input text. Especially, to determine which words are valuable for a particular emotion dimension, an attention layer is trained to weight the words in an input sequence. Furthermore, adversarial training is employed between two attention layers to learn better word weights via a discriminator. In particular, a shared attention layer is incorporated to learn public word weights between two emotion dimensions. Empirical evaluation on the EMOBANK corpus shows that our approach achieves notable improvements in r-values on both EMOBANK Reader{'}s and Writer{'}s multi-dimensional emotion regression tasks in all domains over the state-of-the-art baselines.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1045/
PDF https://www.aclweb.org/anthology/P19-1045
PWC https://paperswithcode.com/paper/adversarial-attention-modeling-for-multi
Repo
Framework
comments powered by Disqus