January 25, 2020

2825 words 14 mins read

Paper Group NANR 49

Paper Group NANR 49

Is this the end? Two-step tokenization of sentence boundaries. What can we learn from natural and artificial dependency trees. Sentence-Level Evidence Embedding for Claim Verification with Hierarchical Attention Networks. signSGD via Zeroth-Order Oracle. Spot and Learn: A Maximum-Entropy Patch Sampler for Few-Shot Image Classification. What goes in …

Is this the end? Two-step tokenization of sentence boundaries

Title Is this the end? Two-step tokenization of sentence boundaries
Authors Linda Wiechetek, Sjur N{\o}rsteb{\o} Moshagen, Thomas Omma
Abstract
Tasks Tokenization
Published 2019-01-01
URL https://www.aclweb.org/anthology/W19-0312/
PDF https://www.aclweb.org/anthology/W19-0312
PWC https://paperswithcode.com/paper/is-this-the-end-two-step-tokenization-of
Repo
Framework

What can we learn from natural and artificial dependency trees

Title What can we learn from natural and artificial dependency trees
Authors Marine Courtin, Chunxiao Yan
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-7915/
PDF https://www.aclweb.org/anthology/W19-7915
PWC https://paperswithcode.com/paper/what-can-we-learn-from-natural-and-artificial
Repo
Framework

Sentence-Level Evidence Embedding for Claim Verification with Hierarchical Attention Networks

Title Sentence-Level Evidence Embedding for Claim Verification with Hierarchical Attention Networks
Authors Jing Ma, Wei Gao, Shafiq Joty, Kam-Fai Wong
Abstract Claim verification is generally a task of verifying the veracity of a given claim, which is critical to many downstream applications. It is cumbersome and inefficient for human fact-checkers to find consistent pieces of evidence, from which solid verdict could be inferred against the claim. In this paper, we propose a novel end-to-end hierarchical attention network focusing on learning to represent coherent evidence as well as their semantic relatedness with the claim. Our model consists of three main components: 1) A coherence-based attention layer embeds coherent evidence considering the claim and sentences from relevant articles; 2) An entailment-based attention layer attends on sentences that can semantically infer the claim on top of the first attention; and 3) An output layer predicts the verdict based on the embedded evidence. Experimental results on three public benchmark datasets show that our proposed model outperforms a set of state-of-the-art baselines.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1244/
PDF https://www.aclweb.org/anthology/P19-1244
PWC https://paperswithcode.com/paper/sentence-level-evidence-embedding-for-claim
Repo
Framework

signSGD via Zeroth-Order Oracle

Title signSGD via Zeroth-Order Oracle
Authors Sijia Liu, Pin-Yu Chen, Xiangyi Chen, Mingyi Hong
Abstract In this paper, we design and analyze a new zeroth-order (ZO) stochastic optimization algorithm, ZO-signSGD, which enjoys dual advantages of gradient-free operations and signSGD. The latter requires only the sign information of gradient estimates but is able to achieve a comparable or even better convergence speed than SGD-type algorithms. Our study shows that ZO signSGD requires $\sqrt{d}$ times more iterations than signSGD, leading to a convergence rate of $O(\sqrt{d}/\sqrt{T})$ under mild conditions, where $d$ is the number of optimization variables, and $T$ is the number of iterations. In addition, we analyze the effects of different types of gradient estimators on the convergence of ZO-signSGD, and propose two variants of ZO-signSGD that at least achieve $O(\sqrt{d}/\sqrt{T})$ convergence rate. On the application side we explore the connection between ZO-signSGD and black-box adversarial attacks in robust deep learning. Our empirical evaluations on image classification datasets MNIST and CIFAR-10 demonstrate the superior performance of ZO-signSGD on the generation of adversarial examples from black-box neural networks.
Tasks Image Classification, Stochastic Optimization
Published 2019-05-01
URL https://openreview.net/forum?id=BJe-DsC5Fm
PDF https://openreview.net/pdf?id=BJe-DsC5Fm
PWC https://paperswithcode.com/paper/signsgd-via-zeroth-order-oracle
Repo
Framework

Spot and Learn: A Maximum-Entropy Patch Sampler for Few-Shot Image Classification

Title Spot and Learn: A Maximum-Entropy Patch Sampler for Few-Shot Image Classification
Authors Wen-Hsuan Chu, Yu-Jhe Li, Jing-Cheng Chang, Yu-Chiang Frank Wang
Abstract Few-shot learning (FSL) requires one to learn from object categories with a small amount of training data (as novel classes), while the remaining categories (as base classes) contain a sufficient amount of data for training. It is often desirable to transfer knowledge from the base classes and derive dominant features efficiently for the novel samples. In this work, we propose a sampling method that de-correlates an image based on maximum entropy reinforcement learning, and extracts varying sequences of patches on every forward-pass with discriminative information observed. This can be viewed as a form of “learned” data augmentation in the sense that we search for different sequences of patches within an image and performs classification with aggregation of the extracted features, resulting in improved FSL performances. In addition, our positive and negative sampling policies along with a newly defined reward function would favorably improve the effectiveness of our model. Our experiments on two benchmark datasets confirm the effectiveness of our framework and its superiority over recent FSL approaches.
Tasks Data Augmentation, Few-Shot Image Classification, Few-Shot Learning, Image Classification
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Chu_Spot_and_Learn_A_Maximum-Entropy_Patch_Sampler_for_Few-Shot_Image_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Chu_Spot_and_Learn_A_Maximum-Entropy_Patch_Sampler_for_Few-Shot_Image_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/spot-and-learn-a-maximum-entropy-patch
Repo
Framework

What goes into a word: generating image descriptions with top-down spatial knowledge

Title What goes into a word: generating image descriptions with top-down spatial knowledge
Authors Mehdi Ghanimifard, Simon Dobnik
Abstract Generating grounded image descriptions requires associating linguistic units with their corresponding visual clues. A common method is to train a decoder language model with attention mechanism over convolutional visual features. Attention weights align the stratified visual features arranged by their location with tokens, most commonly words, in the target description. However, words such as spatial relations (e.g. \textit{next{\textasciitilde}to} and \textit{under}) are not directly referring to geometric arrangements of pixels but to complex geometric and conceptual representations. The aim of this paper is to evaluate what representations facilitate generating image descriptions with spatial relations and lead to better grounded language generation. In particular, we investigate the contribution of three different representational modalities in generating relational referring expressions: (i) pre-trained convolutional visual features, (ii) different top-down geometric relational knowledge between objects, and (iii) world knowledge captured by contextual embeddings in language models.
Tasks Language Modelling, Text Generation
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-8668/
PDF https://www.aclweb.org/anthology/W19-8668
PWC https://paperswithcode.com/paper/what-goes-into-a-word-generating-image
Repo
Framework

Who Sides with Whom? Towards Computational Construction of Discourse Networks for Political Debates

Title Who Sides with Whom? Towards Computational Construction of Discourse Networks for Political Debates
Authors Sebastian Pad{'o}, Andre Blessing, Nico Blokker, Erenay Dayanik, Sebastian Haunss, Jonas Kuhn
Abstract Understanding the structures of political debates (which actors make what claims) is essential for understanding democratic political decision making. The vision of computational construction of such discourse networks from newspaper reports brings together political science and natural language processing. This paper presents three contributions towards this goal: (a) a requirements analysis, linking the task to knowledge base population; (b) an annotated pilot corpus of migration claims based on German newspaper reports; (c) initial modeling results.
Tasks Decision Making, Knowledge Base Population
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1273/
PDF https://www.aclweb.org/anthology/P19-1273
PWC https://paperswithcode.com/paper/who-sides-with-whom-towards-computational
Repo
Framework

Sequence-To-Sequence Domain Adaptation Network for Robust Text Image Recognition

Title Sequence-To-Sequence Domain Adaptation Network for Robust Text Image Recognition
Authors Yaping Zhang, Shuai Nie, Wenju Liu, Xing Xu, Dongxiang Zhang, Heng Tao Shen
Abstract Domain adaptation has shown promising advances for alleviating domain shift problem. However, recent visual domain adaptation works usually focus on non-sequential object recognition with a global coarse alignment, which is inadequate to transfer effective knowledge for sequence-like text images with variable-length fine-grained character information. In this paper, we develop a Sequence-to-Sequence Domain Adaptation Network (SSDAN) for robust text image recognition, which could exploit unsupervised sequence data by an attention-based sequence encoder-decoder network. In the SSDAN, a gated attention similarity (GAS) unit is introduced to adaptively focus on aligning the distribution of the source and target sequence data in an attended character-level feature space rather than a global coarse alignment. Extensive text recognition experiments show the SSDAN could efficiently transfer sequence knowledge and validate the promising power of the proposed model towards real world applications in various recognition scenarios, including the natural scene text, handwritten text and even mathematical expression recognition.
Tasks Domain Adaptation, Object Recognition
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Zhang_Sequence-To-Sequence_Domain_Adaptation_Network_for_Robust_Text_Image_Recognition_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Zhang_Sequence-To-Sequence_Domain_Adaptation_Network_for_Robust_Text_Image_Recognition_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/sequence-to-sequence-domain-adaptation
Repo
Framework

Separate to Adapt: Open Set Domain Adaptation via Progressive Separation

Title Separate to Adapt: Open Set Domain Adaptation via Progressive Separation
Authors Hong Liu, Zhangjie Cao, Mingsheng Long, Jianmin Wang, Qiang Yang
Abstract Domain adaptation has become a resounding success in leveraging labeled data from a source domain to learn an accurate classifier for an unlabeled target domain. When deployed in the wild, the target domain usually contains unknown classes that are not observed in the source domain. Such setting is termed Open Set Domain Adaptation (OSDA). While several methods have been proposed to address OSDA, none of them takes into account the openness of the target domain, which is measured by the proportion of unknown classes in all target classes. Openness is a critical point in open set domain adaptation and exerts a significant impact on performance. In addition, current work aligns the entire target domain with the source domain without excluding unknown samples, which may give rise to negative transfer due to the mismatch between unknown and known classes. To this end, this paper presents Separate to Adapt (STA), an end-to-end approach to open set domain adaptation. The approach adopts a coarse-to-fine weighting mechanism to progressively separate the samples of unknown and known classes, and simultaneously weigh their importance on feature distribution alignment. Our approach allows openness-robust open set domain adaptation, which can be adaptive to a variety of openness in the target domain. We evaluate STA on several benchmark datasets of various openness levels. Results verify that STA significantly outperforms previous methods.
Tasks Domain Adaptation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Liu_Separate_to_Adapt_Open_Set_Domain_Adaptation_via_Progressive_Separation_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Liu_Separate_to_Adapt_Open_Set_Domain_Adaptation_via_Progressive_Separation_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/separate-to-adapt-open-set-domain-adaptation
Repo
Framework

Spatial-Aware Graph Relation Network for Large-Scale Object Detection

Title Spatial-Aware Graph Relation Network for Large-Scale Object Detection
Authors Hang Xu, Chenhan Jiang, Xiaodan Liang, Zhenguo Li
Abstract How to proper encode high-order object relation in the detection system without any external knowledge? How to leverage the information between co-occurrence and locations of objects for better reasoning? These questions are key challenges towards large-scale object detection system that aims to recognize thousands of objects entangled with complex spatial and semantic relationships nowadays. Distilling key relations that may affect object recognition is crucially important since treating each region separately leads to a big performance drop when facing heavy long-tail data distributions and plenty of confusing categories. Recent works try to encode relation by constructing graphs, e.g. using handcraft linguistic knowledge between classes or implicitly learning a fully-connected graph between regions. However, the handcraft linguistic knowledge cannot be individualized for each image due to the semantic gap between linguistic and visual context while the fully-connected graph is inefficient and noisy by incorporating redundant and distracted relations/edges from irrelevant objects and backgrounds. In this work, we introduce a Spatial-aware Graph Relation Network (SGRN) to adaptive discover and incorporate key semantic and spatial relationships for reasoning over each object. Our method considers the relative location layouts and interactions among which can be easily injected into any detection pipelines to boost the performance. Specifically, our SGRN integrates a graph learner module for learning a interpatable sparse graph structure to encode relevant contextual regions and a spatial graph reasoning module with learnable spatial Gaussian kernels to perform graph inference with spatial awareness. Extensive experiments verify the effectiveness of our method, e.g. achieving around 32% improvement on VG(3000 classes) and 28% on ADE in terms of mAP.
Tasks Object Detection, Object Recognition
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Xu_Spatial-Aware_Graph_Relation_Network_for_Large-Scale_Object_Detection_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Xu_Spatial-Aware_Graph_Relation_Network_for_Large-Scale_Object_Detection_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/spatial-aware-graph-relation-network-for
Repo
Framework

Deconstructing multimodality: visual properties and visual context in human semantic processing

Title Deconstructing multimodality: visual properties and visual context in human semantic processing
Authors Christopher Davis, Luana Bulat, Anita Lilla Vero, Ekaterina Shutova
Abstract Multimodal semantic models that extend linguistic representations with additional perceptual input have proved successful in a range of natural language processing (NLP) tasks. Recent research has successfully used neural methods to automatically create visual representations for words. However, these works have extracted visual features from complete images, and have not examined how different kinds of visual information impact performance. In contrast, we construct multimodal models that differentiate between internal visual properties of the objects and their external visual context. We evaluate the models on the task of decoding brain activity associated with the meanings of nouns, demonstrating their advantage over those based on complete images.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-1013/
PDF https://www.aclweb.org/anthology/S19-1013
PWC https://paperswithcode.com/paper/deconstructing-multimodality-visual
Repo
Framework

Projecting Temporal Properties, Events and Actions

Title Projecting Temporal Properties, Events and Actions
Authors Fern, Tim o
Abstract Temporal notions based on a finite set \textit{A} of properties are represented in strings, on which projections are defined that vary the granularity \textit{A}. The structure of properties in \textit{A} is elaborated to describe statives, events and actions, subject to a distinction in meaning (advocated by Levin and Rappaport Hovav) between what the lexicon prescribes and what a context of use supplies. The projections proposed are deployed as labels for records and record types amenable to finite-state methods.
Tasks
Published 2019-05-01
URL https://www.aclweb.org/anthology/W19-0401/
PDF https://www.aclweb.org/anthology/W19-0401
PWC https://paperswithcode.com/paper/projecting-temporal-properties-events-and
Repo
Framework

A Semantic Annotation Scheme for Quantification

Title A Semantic Annotation Scheme for Quantification
Authors Harry Bunt
Abstract This paper describes in brief the proposal called {}QuantML{'} which was accepted by the International Organisation for Standards (ISO) last February as a starting point for developing a standard for the interoperable annotation of quantification phenomena in natural language, as part of the ISO 24617 Semantic Annotation Framework. The proposal, firmly rooted in the theory of generalised quantifiers, neo-Davidsonian semantics, and DRT, covers a wide range of quantification phenomena. The QuantML scheme consists of (1) an abstract syntax which defines {}annotation structures{'} as triples and other set-theoretic constructs; (b) a compositional semantics of annotation structures; (3) an XML representation of annotation structures.
Tasks
Published 2019-05-01
URL https://www.aclweb.org/anthology/W19-0403/
PDF https://www.aclweb.org/anthology/W19-0403
PWC https://paperswithcode.com/paper/a-semantic-annotation-scheme-for
Repo
Framework

A Practical Dialogue-Act-Driven Conversation Model for Multi-Turn Response Selection

Title A Practical Dialogue-Act-Driven Conversation Model for Multi-Turn Response Selection
Authors Harshit Kumar, Arvind Agarwal, Sachindra Joshi
Abstract Dialogue Acts play an important role in conversation modeling. Research has shown the utility of dialogue acts for the response selection task, however, the underlying assumption is that the dialogue acts are readily available, which is impractical, as dialogue acts are rarely available for new conversations. This paper proposes an end-to-end multi-task model for conversation modeling, which is optimized for two tasks, dialogue act prediction and response selection, with the latter being the task of interest. It proposes a novel way of combining the predicted dialogue acts of context and response with the context (previous utterances) and response (follow-up utterance) in a crossway fashion, such that, it achieves at par performance for the response selection task compared to the model that uses actual dialogue acts. Through experiments on two well known datasets, we demonstrate that the multi-task model not only improves the accuracy of the dialogue act prediction task but also improves the MRR for the response selection task. Also, the cross-stitching of dialogue acts of context and response with the context and response is better than using either one of them individually.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1205/
PDF https://www.aclweb.org/anthology/D19-1205
PWC https://paperswithcode.com/paper/a-practical-dialogue-act-driven-conversation
Repo
Framework

The Dreem Headband as an Alternative to Polysomnography for EEG Signal Acquisition and Sleep Staging

Title The Dreem Headband as an Alternative to Polysomnography for EEG Signal Acquisition and Sleep Staging
Authors Pierrick J. Arnal, Valentin Thorey, Michael E. Ballard, Albert Bou Hernandez, Antoine Guillot, Hugo Jourde, Mason Harris, Mathias Guillard, Pascal Van Beers, Mounir Chennaoui, Fabien Sauvet
Abstract Despite the central role of sleep in our lives and the high prevalence of sleep disorders, sleep is still poorly understood. The development of ambulatory technologies capable of monitoring brain activity during sleep longitudinally is critical to advancing sleep science and facilitating the diagnosis of sleep disorders. We introduced the Dreem headband (DH) as an affordable, comfortable, and user-friendly alternative to polysomnography (PSG). The purpose of this study was to assess the signal acquisition of the DH and the performance of its embedded automatic sleep staging algorithms compared to the gold-standard clinical PSG scored by 5 sleep experts. Thirty-one subjects completed an over-night sleep study at a sleep center while wearing both a PSG and the DH simultaneously. We assessed 1) the EEG signal quality between the DH and the PSG, 2) the heart rate, breathing frequency, and respiration rate variability (RRV) agreement between the DH and the PSG, and 3) the performance of the DH’s automatic sleep staging according to AASM guidelines vs. PSG sleep experts manual scoring. Results demonstrate a strong correlation between the EEG signals acquired by the DH and those from the PSG, and the signals acquired by the DH enable monitoring of alpha (r= 0.71 ± 0.13), beta (r= 0.71 ± 0.18), delta (r = 0.76 ± 0.14), and theta (r = 0.61 ± 0.12) frequencies during sleep. The mean absolute error for heart rate, breathing frequency and RRV was 1.2 ± 0.5 bpm, 0.3 ± 0.2 cpm and 3.2 ± 0.6 %, respectively. Automatic Sleep Staging reached an overall accuracy of 83.5 ± 6.4% (F1 score : 83.8 ± 6.3) for the DH to be compared with an average of 86.4 ± 8.0% (F1 score: 86.3 ± 7.4) for the five sleep experts. These results demonstrate the capacity of the DH to both precisely monitor sleep-related physiological signals and process them accurately into sleep stages. This device paves the way for high-quality, large-scale, longitudinal sleep studies.
Tasks EEG, Sleep Quality, Sleep Stage Detection
Published 2019-06-10
URL https://doi.org/10.1101/662734
PDF https://www.biorxiv.org/content/biorxiv/early/2019/06/10/662734.full-text.pdf
PWC https://paperswithcode.com/paper/the-dreem-headband-as-an-alternative-to
Repo
Framework
comments powered by Disqus