January 24, 2020

2590 words 13 mins read

Paper Group NANR 252

Paper Group NANR 252

GSI-UPM at SemEval-2019 Task 5: Semantic Similarity and Word Embeddings for Multilingual Detection of Hate Speech Against Immigrants and Women on Twitter. HATEMINER at SemEval-2019 Task 5: Hate speech detection against Immigrants and Women in Twitter using a Multinomial Naive Bayes Classifier. A3-108 Machine Translation System for LoResMT 2019. Nor …

GSI-UPM at SemEval-2019 Task 5: Semantic Similarity and Word Embeddings for Multilingual Detection of Hate Speech Against Immigrants and Women on Twitter

Title GSI-UPM at SemEval-2019 Task 5: Semantic Similarity and Word Embeddings for Multilingual Detection of Hate Speech Against Immigrants and Women on Twitter
Authors Diego Benito, Oscar Araque, Carlos A. Iglesias
Abstract This paper describes the GSI-UPM system for SemEval-2019 Task 5, which tackles multilingual detection of hate speech on Twitter. The main contribution of the paper is the use of a method based on word embeddings and semantic similarity combined with traditional paradigms, such as n-grams, TF-IDF and POS. This combination of several features is fine-tuned through ablation tests, demonstrating the usefulness of different features. While our approach outperforms baseline classifiers on different sub-tasks, the best of our submitted runs reached the 5th position on the Spanish sub-task A.
Tasks Semantic Similarity, Semantic Textual Similarity, Word Embeddings
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2070/
PDF https://www.aclweb.org/anthology/S19-2070
PWC https://paperswithcode.com/paper/gsi-upm-at-semeval-2019-task-5-semantic
Repo
Framework

HATEMINER at SemEval-2019 Task 5: Hate speech detection against Immigrants and Women in Twitter using a Multinomial Naive Bayes Classifier

Title HATEMINER at SemEval-2019 Task 5: Hate speech detection against Immigrants and Women in Twitter using a Multinomial Naive Bayes Classifier
Authors Nikhil Chakravartula
Abstract This paper describes our participation in the SemEval 2019 Task 5 - Multilingual Detection of Hate. This task aims to identify hate speech against two specific targets, immigrants and women. We compare and contrast the performance of different word and sentence level embeddings on the state-of-the-art classification algorithms. Our final submission is a Multinomial binarized Naive Bayes model for both the subtasks in the English version.
Tasks Hate Speech Detection
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2071/
PDF https://www.aclweb.org/anthology/S19-2071
PWC https://paperswithcode.com/paper/hateminer-at-semeval-2019-task-5-hate-speech
Repo
Framework

A3-108 Machine Translation System for LoResMT 2019

Title A3-108 Machine Translation System for LoResMT 2019
Authors Saumitra Yadav, V Mujadia, an, Manish Shrivastava
Abstract
Tasks Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-6810/
PDF https://www.aclweb.org/anthology/W19-6810
PWC https://paperswithcode.com/paper/a3-108-machine-translation-system-for-loresmt
Repo
Framework

Normalized Wasserstein for Mixture Distributions With Applications in Adversarial Learning and Domain Adaptation

Title Normalized Wasserstein for Mixture Distributions With Applications in Adversarial Learning and Domain Adaptation
Authors Yogesh Balaji, Rama Chellappa, Soheil Feizi
Abstract Understanding proper distance measures between distributions is at the core of several learning tasks such as generative models, domain adaptation, clustering, etc. In this work, we focus on mixture distributions that arise naturally in several application domains where the data contains different sub-populations. For mixture distributions, established distance measures such as the Wasserstein distance do not take into account imbalanced mixture proportions. Thus, even if two mixture distributions have identical mixture components but different mixture proportions, the Wasserstein distance between them will be large. This often leads to undesired results in distance-based learning methods for mixture distributions. In this paper, we resolve this issue by introducing the Normalized Wasserstein measure. The key idea is to introduce mixture proportions as optimization variables, effectively normalizing mixture proportions in the Wasserstein formulation. Using the proposed normalized Wasserstein measure leads to significant performance gains for mixture distributions with imbalanced mixture proportions compared to the vanilla Wasserstein distance. We demonstrate the effectiveness of the proposed measure in GANs, domain adaptation and adversarial clustering in several benchmark datasets.
Tasks Domain Adaptation
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Balaji_Normalized_Wasserstein_for_Mixture_Distributions_With_Applications_in_Adversarial_Learning_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Balaji_Normalized_Wasserstein_for_Mixture_Distributions_With_Applications_in_Adversarial_Learning_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/normalized-wasserstein-for-mixture
Repo
Framework

MineriaUNAM at SemEval-2019 Task 5: Detecting Hate Speech in Twitter using Multiple Features in a Combinatorial Framework

Title MineriaUNAM at SemEval-2019 Task 5: Detecting Hate Speech in Twitter using Multiple Features in a Combinatorial Framework
Authors Luis Enrique Argota Vega, Jorge Carlos Reyes-Maga{~n}a, Helena G{'o}mez-Adorno, Gemma Bel-Enguix
Abstract This paper presents our approach to the Task 5 of Semeval-2019, which aims at detecting hate speech against immigrants and women in Twitter. The task consists of two sub-tasks, in Spanish and English: (A) detection of hate speech and (B) classification of hateful tweets as aggressive or not, and identification of the target harassed as individual or group. We used linguistically motivated features and several types of n-grams (words, characters, functional words, punctuation symbols, POS, among others). For task A, we trained a Support Vector Machine using a combinatorial framework, whereas for task B we followed a multi-labeled approach using the Random Forest classifier. Our approach achieved the highest F1-score in sub-task A for the Spanish language.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2079/
PDF https://www.aclweb.org/anthology/S19-2079
PWC https://paperswithcode.com/paper/mineriaunam-at-semeval-2019-task-5-detecting
Repo
Framework

A large-scale study of the effects of word frequency and predictability in naturalistic reading

Title A large-scale study of the effects of word frequency and predictability in naturalistic reading
Authors Cory Shain
Abstract A number of psycholinguistic studies have factorially manipulated words{'} contextual predictabilities and corpus frequencies and shown separable effects of each on measures of human sentence processing, a pattern which has been used to support distinct mechanisms underlying prediction on the one hand and lexical retrieval on the other. This paper examines the generalizability of this finding to more realistic conditions of sentence processing by studying effects of frequency and predictability in three large-scale naturalistic reading corpora. Results show significant effects of word frequency and predictability in isolation but no effect of frequency over and above predictability, and thus do not provide evidence of distinct mechanisms. The non-replication of separable effects in a naturalistic setting raises doubts about the existence of such a distinction in everyday sentence comprehension. Instead, these results are consistent with previous claims that apparent effects of frequency are underlyingly effects of predictability.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-1413/
PDF https://www.aclweb.org/anthology/N19-1413
PWC https://paperswithcode.com/paper/a-large-scale-study-of-the-effects-of-word
Repo
Framework

Negative Focus Detection via Contextual Attention Mechanism

Title Negative Focus Detection via Contextual Attention Mechanism
Authors Longxiang Shen, Bowei Zou, Yu Hong, Guodong Zhou, Qiaoming Zhu, AiTi Aw
Abstract Negation is a universal but complicated linguistic phenomenon, which has received considerable attention from the NLP community over the last decade, since a negated statement often carries both an explicit negative focus and implicit positive meanings. For the sake of understanding a negated statement, it is critical to precisely detect the negative focus in context. However, how to capture contextual information for negative focus detection is still an open challenge. To well address this, we come up with an attention-based neural network to model contextual information. In particular, we introduce a framework which consists of a Bidirectional Long Short-Term Memory (BiLSTM) neural network and a Conditional Random Fields (CRF) layer to effectively encode the order information and the long-range context dependency in a sentence. Moreover, we design two types of attention mechanisms, word-level contextual attention and topic-level contextual attention, to take advantage of contextual information across sentences from both the word perspective and the topic perspective, respectively. Experimental results on the SEM{'}12 shared task corpus show that our approach achieves the best performance on negative focus detection, yielding an absolute improvement of 2.11{%} over the state-of-the-art. This demonstrates the great effectiveness of the two types of contextual attention mechanisms.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1230/
PDF https://www.aclweb.org/anthology/D19-1230
PWC https://paperswithcode.com/paper/negative-focus-detection-via-contextual
Repo
Framework

Spatiotemporal Feature Residual Propagation for Action Prediction

Title Spatiotemporal Feature Residual Propagation for Action Prediction
Authors He Zhao, Richard P. Wildes
Abstract Recognizing actions from limited preliminary video observations has seen considerable recent progress. Typically, however, such progress has been had without explicitly modeling fine-grained motion evolution as a potentially valuable information source. In this study, we address this task by investigating how action patterns evolve over time in a spatial feature space. There are three key components to our system. First, we work with intermediate-layer ConvNet features, which allow for abstraction from raw data, while retaining spatial layout, which is sacrificed in approaches that rely on vectorized global representations. Second, instead of propagating features per se, we propagate their residuals across time, which allows for a compact representation that reduces redundancy while retaining essential information about evolution over time. Third, we employ a Kalman filter to combat error build-up and unify across prediction start times. Extensive experimental results on the JHMDB21, UCF101 and BIT datasets show that our approach leads to a new state-of-the-art in action prediction.
Tasks
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Zhao_Spatiotemporal_Feature_Residual_Propagation_for_Action_Prediction_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Zhao_Spatiotemporal_Feature_Residual_Propagation_for_Action_Prediction_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/spatiotemporal-feature-residual-propagation
Repo
Framework

Attribute Manipulation Generative Adversarial Networks for Fashion Images

Title Attribute Manipulation Generative Adversarial Networks for Fashion Images
Authors Kenan E. Ak, Joo Hwee Lim, Jo Yew Tham, Ashraf A. Kassim
Abstract Recent advances in Generative Adversarial Networks (GANs) have made it possible to conduct multi-domain image-to-image translation using a single generative network. While recent methods such as Ganimation and SaGAN are able to conduct translations on attribute-relevant regions using attention, they do not perform well when the number of attributes increases as the training of attention masks mostly rely on classification losses. To address this and other limitations, we introduce Attribute Manipulation Generative Adversarial Networks (AMGAN) for fashion images. While AMGAN’s generator network uses class activation maps (CAMs) to empower its attention mechanism, it also exploits perceptual losses by assigning reference (target) images based on attribute similarities. AMGAN incorporates an additional discriminator network that focuses on attribute-relevant regions to detect unrealistic translations. Additionally, AMGAN can be controlled to perform attribute manipulations on specific regions such as the sleeve or torso regions. Experiments show that AMGAN outperforms state-of-the-art methods using traditional evaluation metrics as well as an alternative one that is based on image retrieval.
Tasks Image Retrieval, Image-to-Image Translation
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Ak_Attribute_Manipulation_Generative_Adversarial_Networks_for_Fashion_Images_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Ak_Attribute_Manipulation_Generative_Adversarial_Networks_for_Fashion_Images_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/attribute-manipulation-generative-adversarial
Repo
Framework

INGEOTEC at SemEval-2019 Task 5 and Task 6: A Genetic Programming Approach for Text Classification

Title INGEOTEC at SemEval-2019 Task 5 and Task 6: A Genetic Programming Approach for Text Classification
Authors Mario Graff, Mir, Sabino a-Jim{'e}nez, Eric Tellez, Daniela Alej Ochoa, ra
Abstract This paper describes our participation in HatEval and OffensEval challenges for English and Spanish languages. We used several approaches, B4MSA, FastText, and EvoMSA. Best results were achieved with EvoMSA, which is a multilingual and domain-independent architecture that combines the prediction of different knowledge sources to solve text classification problems.
Tasks Text Classification
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2114/
PDF https://www.aclweb.org/anthology/S19-2114
PWC https://paperswithcode.com/paper/ingeotec-at-semeval-2019-task-5-and-task-6-a
Repo
Framework

Towards Natural and Accurate Future Motion Prediction of Humans and Animals

Title Towards Natural and Accurate Future Motion Prediction of Humans and Animals
Authors Zhenguang Liu, Shuang Wu, Shuyuan Jin, Qi Liu, Shijian Lu, Roger Zimmermann, Li Cheng
Abstract Anticipating the future motions of 3D articulate objects is challenging due to its non-linear and highly stochastic nature. Current approaches typically represent the skeleton of an articulate object as a set of 3D joints, which unfortunately ignores the relationship between joints, and fails to encode fine-grained anatomical constraints. Moreover, conventional recurrent neural networks, such as LSTM and GRU, are employed to model motion contexts, which inherently have difficulties in capturing long-term dependencies. To address these problems, we propose to explicitly encode anatomical constraints by modeling their skeletons with a Lie algebra representation. Importantly, a hierarchical recurrent network structure is developed to simultaneously encodes local contexts of individual frames and global contexts of the sequence. We proceed to explore the applications of our approach to several distinct quantities including human, fish, and mouse. Extensive experiments show that our approach achieves more natural and accurate predictions over state-of-the-art methods.
Tasks motion prediction
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Liu_Towards_Natural_and_Accurate_Future_Motion_Prediction_of_Humans_and_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Liu_Towards_Natural_and_Accurate_Future_Motion_Prediction_of_Humans_and_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/towards-natural-and-accurate-future-motion
Repo
Framework

Multi-task Learning for Natural Language Generation in Task-Oriented Dialogue

Title Multi-task Learning for Natural Language Generation in Task-Oriented Dialogue
Authors Chenguang Zhu, Michael Zeng, Xuedong Huang
Abstract In task-oriented dialogues, Natural Language Generation (NLG) is the final yet crucial step to produce user-facing system utterances. The result of NLG is directly related to the perceived quality and usability of a dialogue system. While most existing systems provide semantically correct responses given goals to present, they struggle to match the variation and fluency in the human language. In this paper, we propose a novel multi-task learning framework, NLG-LM, for natural language generation. In addition to generating high-quality responses conveying the required information, it also explicitly targets for naturalness in generated responses via an unconditioned language model. This can significantly improve the learning of style and variation in human language. Empirical results show that this multi-task learning framework outperforms previous models across multiple datasets. For example, it improves the previous best BLEU score on the E2E-NLG dataset by 2.2{%}, and on the Laptop dataset by 6.1{%}.
Tasks Language Modelling, Multi-Task Learning, Text Generation
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1123/
PDF https://www.aclweb.org/anthology/D19-1123
PWC https://paperswithcode.com/paper/multi-task-learning-for-natural-language
Repo
Framework

Duluth at SemEval-2019 Task 4: The Pioquinto Manterola Hyperpartisan News Detector

Title Duluth at SemEval-2019 Task 4: The Pioquinto Manterola Hyperpartisan News Detector
Authors Saptarshi Sengupta, Ted Pedersen
Abstract This paper describes the Pioquinto Manterola Hyperpartisan News Detector, which participated in SemEval-2019 Task 4. Hyperpartisan news is highly polarized and takes a very biased or one{–}sided view of a particular story. We developed two variants of our system, the more successful was a Logistic Regression classifier based on unigram features. This was our official entry in the task, and it placed 23rd of 42 participating teams. Our second variant was a Convolutional Neural Network that did not perform as well.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2162/
PDF https://www.aclweb.org/anthology/S19-2162
PWC https://paperswithcode.com/paper/duluth-at-semeval-2019-task-4-the-pioquinto
Repo
Framework

Enhancing Neural Data-To-Text Generation Models with External Background Knowledge

Title Enhancing Neural Data-To-Text Generation Models with External Background Knowledge
Authors Shuang Chen, Jinpeng Wang, Xiaocheng Feng, Feng Jiang, Bing Qin, Chin-Yew Lin
Abstract Recent neural models for data-to-text generation rely on massive parallel pairs of data and text to learn the writing knowledge. They often assume that writing knowledge can be acquired from the training data alone. However, when people are writing, they not only rely on the data but also consider related knowledge. In this paper, we enhance neural data-to-text models with external knowledge in a simple but effective way to improve the fidelity of generated text. Besides relying on parallel data and text as in previous work, our model attends to relevant external knowledge, encoded as a temporary memory, and combines this knowledge with the context representation of data before generating words. This allows the model to infer relevant facts which are not explicitly stated in the data table from an external knowledge source. Experimental results on twenty-one Wikipedia infobox-to-text datasets show our model, KBAtt, consistently improves a state-of-the-art model on most of the datasets. In addition, to quantify when and why external knowledge is effective, we design a metric, KBGain, which shows a strong correlation with the observed performance boost. This result demonstrates the relevance of external knowledge and sparseness of original data are the main factors affecting system performance.
Tasks Data-to-Text Generation, Text Generation
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1299/
PDF https://www.aclweb.org/anthology/D19-1299
PWC https://paperswithcode.com/paper/enhancing-neural-data-to-text-generation
Repo
Framework

UNBNLP at SemEval-2019 Task 5 and 6: Using Language Models to Detect Hate Speech and Offensive Language

Title UNBNLP at SemEval-2019 Task 5 and 6: Using Language Models to Detect Hate Speech and Offensive Language
Authors Ali Hakimi Parizi, Milton King, Paul Cook
Abstract In this paper we apply a range of approaches to language modeling {–} including word-level n-gram and neural language models, and character-level neural language models {–} to the problem of detecting hate speech and offensive language. Our findings indicate that language models are able to capture knowledge of whether text is hateful or offensive. However, our findings also indicate that more conventional approaches to text classification often perform similarly or better.
Tasks Language Modelling, Text Classification
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2092/
PDF https://www.aclweb.org/anthology/S19-2092
PWC https://paperswithcode.com/paper/unbnlp-at-semeval-2019-task-5-and-6-using
Repo
Framework
comments powered by Disqus