January 24, 2020

2225 words 11 mins read

Paper Group NANR 175

Paper Group NANR 175

Generating Responses with a Specific Emotion in Dialog. SemEval-2019 Task 4: Hyperpartisan News Detection. Re-randomized Densification for One Permutation Hashing and Bin-wise Consistent Weighted Sampling. SimpleNLG-DE: Adapting SimpleNLG 4 to German. ZigZagNet: Fusing Top-Down and Bottom-Up Context for Object Segmentation. Machine Reading Comprehe …

Generating Responses with a Specific Emotion in Dialog

Title Generating Responses with a Specific Emotion in Dialog
Authors Zhenqiao Song, Xiaoqing Zheng, Lu Liu, Mu Xu, Xuanjing Huang
Abstract It is desirable for dialog systems to have capability to express specific emotions during a conversation, which has a direct, quantifiable impact on improvement of their usability and user satisfaction. After a careful investigation of real-life conversation data, we found that there are at least two ways to express emotions with language. One is to describe emotional states by explicitly using strong emotional words; another is to increase the intensity of the emotional experiences by implicitly combining neutral words in distinct ways. We propose an emotional dialogue system (EmoDS) that can generate the meaningful responses with a coherent structure for a post, and meanwhile express the desired emotion explicitly or implicitly within a unified framework. Experimental results showed EmoDS performed better than the baselines in BLEU, diversity and the quality of emotional expression.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1359/
PDF https://www.aclweb.org/anthology/P19-1359
PWC https://paperswithcode.com/paper/generating-responses-with-a-specific-emotion
Repo
Framework

SemEval-2019 Task 4: Hyperpartisan News Detection

Title SemEval-2019 Task 4: Hyperpartisan News Detection
Authors Johannes Kiesel, Maria Mestre, Rishabh Shukla, Emmanuel Vincent, Payam Adineh, David Corney, Benno Stein, Martin Potthast
Abstract Hyperpartisan news is news that takes an extreme left-wing or right-wing standpoint. If one is able to reliably compute this meta information, news articles may be automatically tagged, this way encouraging or discouraging readers to consume the text. It is an open question how successfully hyperpartisan news detection can be automated, and the goal of this SemEval task was to shed light on the state of the art. We developed new resources for this purpose, including a manually labeled dataset with 1,273 articles, and a second dataset with 754,000 articles, labeled via distant supervision. The interest of the research community in our task exceeded all our expectations: The datasets were downloaded about 1,000 times, 322 teams registered, of which 184 configured a virtual machine on our shared task cloud service TIRA, of which in turn 42 teams submitted a valid run. The best team achieved an accuracy of 0.822 on a balanced sample (yes : no hyperpartisan) drawn from the manually tagged corpus; an ensemble of the submitted systems increased the accuracy by 0.048.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2145/
PDF https://www.aclweb.org/anthology/S19-2145
PWC https://paperswithcode.com/paper/semeval-2019-task-4-hyperpartisan-news
Repo
Framework

Re-randomized Densification for One Permutation Hashing and Bin-wise Consistent Weighted Sampling

Title Re-randomized Densification for One Permutation Hashing and Bin-wise Consistent Weighted Sampling
Authors Ping Li, Xiaoyun Li, Cun-Hui Zhang
Abstract Jaccard similarity is widely used as a distance measure in many machine learning and search applications. Typically, hashing methods are essential for the use of Jaccard similarity to be practical in large-scale settings. For hashing binary (0/1) data, the idea of one permutation hashing (OPH) with densification significantly accelerates traditional minwise hashing algorithms while providing unbiased and accurate estimates. In this paper, we propose a strategy named “re-randomization” in the process of densification that could achieve the smallest variance among all densification schemes. The success of this idea naturally inspires us to generalize one permutation hashing to weighted (non-binary) data, which results in the socalled “bin-wise consistent weighted sampling (BCWS)” algorithm. We analyze the behavior of BCWS and compare it with a recent alternative. Extensive experiments on various datasets illustrates the effectiveness of our proposed methods.
Tasks
Published 2019-12-01
URL http://papers.nips.cc/paper/9721-re-randomized-densification-for-one-permutation-hashing-and-bin-wise-consistent-weighted-sampling
PDF http://papers.nips.cc/paper/9721-re-randomized-densification-for-one-permutation-hashing-and-bin-wise-consistent-weighted-sampling.pdf
PWC https://paperswithcode.com/paper/re-randomized-densification-for-one
Repo
Framework

SimpleNLG-DE: Adapting SimpleNLG 4 to German

Title SimpleNLG-DE: Adapting SimpleNLG 4 to German
Authors Daniel Braun, Kira Klimt, Daniela Schneider, Florian Matthes
Abstract SimpleNLG is a popular open source surface realiser for the English language. For German, however, the availability of open source and non-domain specific realisers is sparse, partly due to the complexity of the German language. In this paper, we present SimpleNLG-DE, an adaption of SimpleNLG to German. We discuss which parts of the German language have been implemented and how we evaluated our implementation using the TIGER Corpus and newly created data-sets.
Tasks
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-8651/
PDF https://www.aclweb.org/anthology/W19-8651
PWC https://paperswithcode.com/paper/simplenlg-de-adapting-simplenlg-4-to-german
Repo
Framework

ZigZagNet: Fusing Top-Down and Bottom-Up Context for Object Segmentation

Title ZigZagNet: Fusing Top-Down and Bottom-Up Context for Object Segmentation
Authors Di Lin, Dingguo Shen, Siting Shen, Yuanfeng Ji, Dani Lischinski, Daniel Cohen-Or, Hui Huang
Abstract Multi-scale context information has proven to be essential for object segmentation tasks. Recent works construct the multi-scale context by aggregating convolutional feature maps extracted by different levels of a deep neural network. This is typically done by propagating and fusing features in a one-directional, top-down and bottom-up, manner. In this work, we introduce ZigZagNet, which aggregates a richer multi-context feature map by using not only dense top-down and bottom-up propagation, but also by introducing pathways crossing between different levels of the top-down and the bottom-up hierarchies, in a zig-zag fashion. Furthermore, the context information is exchanged and aggregated over multiple stages, where the fused feature maps from one stage are fed into the next one, yielding a more comprehensive context for improved segmentation performance. Our extensive evaluation on the public benchmarks demonstrates that ZigZagNet surpasses the state-of-the-art accuracy for both semantic segmentation and instance segmentation tasks.
Tasks Instance Segmentation, Semantic Segmentation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Lin_ZigZagNet_Fusing_Top-Down_and_Bottom-Up_Context_for_Object_Segmentation_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Lin_ZigZagNet_Fusing_Top-Down_and_Bottom-Up_Context_for_Object_Segmentation_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/zigzagnet-fusing-top-down-and-bottom-up
Repo
Framework

Machine Reading Comprehension Using Structural Knowledge Graph-aware Network

Title Machine Reading Comprehension Using Structural Knowledge Graph-aware Network
Authors Delai Qiu, Yuanzhe Zhang, Xinwei Feng, Xiangwen Liao, Wenbin Jiang, Yajuan Lyu, Kang Liu, Jun Zhao
Abstract Leveraging external knowledge is an emerging trend in machine comprehension task. Previous work usually utilizes knowledge graphs such as ConceptNet as external knowledge, and extracts triples from them to enhance the initial representation of the machine comprehension context. However, such method cannot capture the structural information in the knowledge graph. To this end, we propose a Structural Knowledge Graph-aware Network(SKG) model, constructing sub-graphs for entities in the machine comprehension context. Our method dynamically updates the representation of the knowledge according to the structural information of the constructed sub-graph. Experiments show that SKG achieves state-of-the-art performance on the ReCoRD dataset.
Tasks Knowledge Graphs, Machine Reading Comprehension, Reading Comprehension
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1602/
PDF https://www.aclweb.org/anthology/D19-1602
PWC https://paperswithcode.com/paper/machine-reading-comprehension-using
Repo
Framework

Predicting the Difficulty of Multiple Choice Questions in a High-stakes Medical Exam

Title Predicting the Difficulty of Multiple Choice Questions in a High-stakes Medical Exam
Authors Le An Ha, Victoria Yaneva, Peter Baldwin, Janet Mee
Abstract Predicting the construct-relevant difficulty of Multiple-Choice Questions (MCQs) has the potential to reduce cost while maintaining the quality of high-stakes exams. In this paper, we propose a method for estimating the difficulty of MCQs from a high-stakes medical exam, where all questions were deliberately written to a common reading level. To accomplish this, we extract a large number of linguistic features and embedding types, as well as features quantifying the difficulty of the items for an automatic question-answering system. The results show that the proposed approach outperforms various baselines with a statistically significant difference. Best results were achieved when using the full feature set, where embeddings had the highest predictive power, followed by linguistic features. An ablation study of the various types of linguistic features suggested that information from all levels of linguistic processing contributes to predicting item difficulty, with features related to semantic ambiguity and the psycholinguistic properties of words having a slightly higher importance. Owing to its generic nature, the presented approach has the potential to generalize over other exams containing MCQs.
Tasks Question Answering
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4402/
PDF https://www.aclweb.org/anthology/W19-4402
PWC https://paperswithcode.com/paper/predicting-the-difficulty-of-multiple-choice
Repo
Framework

The Anisotropic Noise in Stochastic Gradient Descent: Its Behavior of Escaping from Minima and Regularization Effects

Title The Anisotropic Noise in Stochastic Gradient Descent: Its Behavior of Escaping from Minima and Regularization Effects
Authors Zhanxing Zhu, Jingfeng Wu, Bing Yu, Lei Wu, Jinwen Ma
Abstract Understanding the behavior of stochastic gradient descent (SGD) in the context of deep neural networks has raised lots of concerns recently. Along this line, we theoretically study a general form of gradient based optimization dynamics with unbiased noise, which unifies SGD and standard Langevin dynamics. Through investigating this general optimization dynamics, we analyze the behavior of SGD on escaping from minima and its regularization effects. A novel indicator is derived to characterize the efficiency of escaping from minima through measuring the alignment of noise covariance and the curvature of loss function. Based on this indicator, two conditions are established to show which type of noise structure is superior to isotropic noise in term of escaping efficiency. We further show that the anisotropic noise in SGD satisfies the two conditions, and thus helps to escape from sharp and poor minima effectively, towards more stable and flat minima that typically generalize well. We verify our understanding through comparing this anisotropic diffusion with full gradient descent plus isotropic diffusion (i.e. Langevin dynamics) and other types of position-dependent noise.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=H1M7soActX
PDF https://openreview.net/pdf?id=H1M7soActX
PWC https://paperswithcode.com/paper/the-anisotropic-noise-in-stochastic-gradient-1
Repo
Framework

Co-Mining: Deep Face Recognition With Noisy Labels

Title Co-Mining: Deep Face Recognition With Noisy Labels
Authors Xiaobo Wang, Shuo Wang, Jun Wang, Hailin Shi, Tao Mei
Abstract Face recognition has achieved significant progress with the growing scale of collected datasets, which empowers us to train strong convolutional neural networks (CNNs). While a variety of CNN architectures and loss functions have been devised recently, we still have a limited understanding of how to train the CNN models with the label noise inherent in existing face recognition datasets. To address this issue, this paper develops a novel co-mining strategy to effectively train on the datasets with noisy labels. Specifically, we simultaneously use the loss values as the cue to detect noisy labels, exchange the high-confidence clean faces to alleviate the errors accumulated issue caused by the sample-selection bias, and re-weight the predicted clean faces to make them dominate the discriminative model training in a mini-batch fashion. Extensive experiments by training on three popular datasets (i.e., CASIA-WebFace, MS-Celeb-1M and VggFace2) and testing on several benchmarks, including LFW, AgeDB, CFP, CALFW, CPLFW, RFW, and MegaFace, have demonstrated the effectiveness of our new approach over the state-of-the-art alternatives.
Tasks Face Recognition
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Wang_Co-Mining_Deep_Face_Recognition_With_Noisy_Labels_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Wang_Co-Mining_Deep_Face_Recognition_With_Noisy_Labels_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/co-mining-deep-face-recognition-with-noisy
Repo
Framework

NUIG at the FinSBD Task: Sentence Boundary Detection for Noisy Financial PDFs in English and French

Title NUIG at the FinSBD Task: Sentence Boundary Detection for Noisy Financial PDFs in English and French
Authors Tobias Daudert, Sina Ahmadi
Abstract
Tasks Boundary Detection
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5519/
PDF https://www.aclweb.org/anthology/W19-5519
PWC https://paperswithcode.com/paper/nuig-at-the-finsbd-task-sentence-boundary
Repo
Framework

Recurrent Experience Replay in Distributed Reinforcement Learning

Title Recurrent Experience Replay in Distributed Reinforcement Learning
Authors Steven Kapturowski, Georg Ostrovski, Will Dabney, John Quan, Remi Munos
Abstract Building on the recent successes of distributed training of RL agents, in this paper we investigate the training of RNN-based RL agents from distributed prioritized experience replay. We study the effects of parameter lag resulting in representational drift and recurrent state staleness and empirically derive an improved training strategy. Using a single network architecture and fixed set of hyperparameters, the resulting agent, Recurrent Replay Distributed DQN, quadruples the previous state of the art on Atari-57, and surpasses the state of the art on DMLab-30. It is the first agent to exceed human-level performance in 52 of the 57 Atari games.
Tasks Atari Games
Published 2019-05-01
URL https://openreview.net/forum?id=r1lyTjAqYX
PDF https://openreview.net/pdf?id=r1lyTjAqYX
PWC https://paperswithcode.com/paper/recurrent-experience-replay-in-distributed
Repo
Framework

Proceedings of the 6th International Sanskrit Computational Linguistics Symposium

Title Proceedings of the 6th International Sanskrit Computational Linguistics Symposium
Authors
Abstract
Tasks
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-7500/
PDF https://www.aclweb.org/anthology/W19-7500
PWC https://paperswithcode.com/paper/proceedings-of-the-6th-international-sanskrit
Repo
Framework

Future Directions in Technological Support for Language Documentation

Title Future Directions in Technological Support for Language Documentation
Authors Daan van Esch, Ben Foley, Nay San
Abstract
Tasks
Published 2019-02-01
URL https://www.aclweb.org/anthology/W19-6003/
PDF https://www.aclweb.org/anthology/W19-6003
PWC https://paperswithcode.com/paper/future-directions-in-technological-support
Repo
Framework

Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Industry Papers)

Title Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Industry Papers)
Authors
Abstract
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-2000/
PDF https://www.aclweb.org/anthology/N19-2000
PWC https://paperswithcode.com/paper/proceedings-of-the-2019-conference-of-the-1
Repo
Framework

No-Press Diplomacy: Modeling Multi-Agent Gameplay

Title No-Press Diplomacy: Modeling Multi-Agent Gameplay
Authors Philip Paquette, Yuchen Lu, Seton Steven Bocco, Max Smith, Satya O.-G., Jonathan K. Kummerfeld, Joelle Pineau, Satinder Singh, Aaron C. Courville
Abstract Diplomacy is a seven-player non-stochastic, non-cooperative game, where agents acquire resources through a mix of teamwork and betrayal. Reliance on trust and coordination makes Diplomacy the first non-cooperative multi-agent benchmark for complex sequential social dilemmas in a rich environment. In this work, we focus on training an agent that learns to play the No Press version of Diplomacy where there is no dedicated communication channel between players. We present DipNet, a neural-network-based policy model for No Press Diplomacy. The model was trained on a new dataset of more than 150,000 human games. Our model is trained by supervised learning (SL) from expert trajectories, which is then used to initialize a reinforcement learning (RL) agent trained through self-play. Both the SL and the RL agent demonstrate state-of-the-art No Press performance by beating popular rule-based bots.
Tasks
Published 2019-12-01
URL http://papers.nips.cc/paper/8697-no-press-diplomacy-modeling-multi-agent-gameplay
PDF http://papers.nips.cc/paper/8697-no-press-diplomacy-modeling-multi-agent-gameplay.pdf
PWC https://paperswithcode.com/paper/no-press-diplomacy-modeling-multi-agent-1
Repo
Framework
comments powered by Disqus