January 24, 2020

2839 words 14 mins read

Paper Group NANR 240

Paper Group NANR 240

StrokeNet: A Neural Painting Environment. Sample Efficient Imitation Learning for Continuous Control. Amplifying the Neural Power Spectrum. Exploring Graph-Algebraic CCG Combinators for Syntactic-Semantic AMR Parsing. What a difference a pixel makes: An empirical examination of features used by CNNs for categorisation. Joint extraction of entities …

StrokeNet: A Neural Painting Environment

Title StrokeNet: A Neural Painting Environment
Authors Ningyuan Zheng, Yifan Jiang, Dingjiang Huang
Abstract We’ve seen tremendous success of image generating models these years. Generating images through a neural network is usually pixel-based, which is fundamentally different from how humans create artwork using brushes. To imitate human drawing, interactions between the environment and the agent is required to allow trials. However, the environment is usually non-differentiable, leading to slow convergence and massive computation. In this paper we try to address the discrete nature of software environment with an intermediate, differentiable simulation. We present StrokeNet, a novel model where the agent is trained upon a well-crafted neural approximation of the painting environment. With this approach, our agent was able to learn to write characters such as MNIST digits faster than reinforcement learning approaches in an unsupervised manner. Our primary contribution is the neural simulation of a real-world environment. Furthermore, the agent trained with the emulated environment is able to directly transfer its skills to real-world software.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=HJxwDiActX
PDF https://openreview.net/pdf?id=HJxwDiActX
PWC https://paperswithcode.com/paper/strokenet-a-neural-painting-environment
Repo
Framework

Sample Efficient Imitation Learning for Continuous Control

Title Sample Efficient Imitation Learning for Continuous Control
Authors Fumihiro Sasaki, Tetsuya Yohira, Atsuo Kawaguchi
Abstract The goal of imitation learning (IL) is to enable a learner to imitate expert behavior given expert demonstrations. Recently, generative adversarial imitation learning (GAIL) has shown significant progress on IL for complex continuous tasks. However, GAIL and its extensions require a large number of environment interactions during training. In real-world environments, the more an IL method requires the learner to interact with the environment for better imitation, the more training time it requires, and the more damage it causes to the environments and the learner itself. We believe that IL algorithms could be more applicable to real-world problems if the number of interactions could be reduced. In this paper, we propose a model-free IL algorithm for continuous control. Our algorithm is made up mainly three changes to the existing adversarial imitation learning (AIL) methods – (a) adopting off-policy actor-critic (Off-PAC) algorithm to optimize the learner policy, (b) estimating the state-action value using off-policy samples without learning reward functions, and (c) representing the stochastic policy function so that its outputs are bounded. Experimental results show that our algorithm achieves competitive results with GAIL while significantly reducing the environment interactions.
Tasks Continuous Control, Imitation Learning
Published 2019-05-01
URL https://openreview.net/forum?id=BkN5UoAqF7
PDF https://openreview.net/pdf?id=BkN5UoAqF7
PWC https://paperswithcode.com/paper/sample-efficient-imitation-learning-for
Repo
Framework

Amplifying the Neural Power Spectrum

Title Amplifying the Neural Power Spectrum
Authors J. Andrew Doyle, Paule-Joanne Toussaint, Alan C. Evans
Abstract We introduce a novel method that employs a parametric model of human electroen-cephalographic (EEG) brain signal power spectra to evaluate cognitive science experiments and test scientific hypotheses. We develop the Neural Power Amplifier (NPA), a data-driven approach to EEG pre-processing that can replace current filtering strategies with a principled method based on combining filters with log-arithmic and Gaussian magnitude responses. Presenting the first time domain evidence to validate an increasingly popular model for neural power spectra, we show that filtering out the 1/f background signal and selecting peaks improves a time-domain decoding experiment for visual stimulus of human faces versus random noise.
Tasks EEG
Published 2019-06-04
URL https://doi.org/10.1101/659268
PDF https://www.biorxiv.org/content/biorxiv/early/2019/06/04/659268.full.pdf
PWC https://paperswithcode.com/paper/amplifying-the-neural-power-spectrum
Repo
Framework

Exploring Graph-Algebraic CCG Combinators for Syntactic-Semantic AMR Parsing

Title Exploring Graph-Algebraic CCG Combinators for Syntactic-Semantic AMR Parsing
Authors Sebastian Beschke
Abstract We describe a new approach to semantic parsing based on Combinatory Categorial Grammar (CCG). The grammar{'}s semantic construction operators are defined in terms of a graph algebra, which allows our system to induce a compact CCG lexicon. We introduce an expectation maximisation algorithm which we use to filter our lexicon down to 2500 lexical templates. Our system achieves a semantic triple (Smatch) precision that is competitive with other CCG-based AMR parsing approaches.
Tasks Amr Parsing, Semantic Parsing
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-1014/
PDF https://www.aclweb.org/anthology/R19-1014
PWC https://paperswithcode.com/paper/exploring-graph-algebraic-ccg-combinators-for
Repo
Framework

What a difference a pixel makes: An empirical examination of features used by CNNs for categorisation

Title What a difference a pixel makes: An empirical examination of features used by CNNs for categorisation
Authors Gaurav Malhotra, Jeffrey Bowers
Abstract Convolutional neural networks (CNNs) were inspired by human vision and, in some settings, achieve a performance comparable to human object recognition. This has lead to the speculation that both systems use similar mechanisms to perform recognition. In this study, we conducted a series of simulations that indicate that there is a fundamental difference between human vision and CNNs: while object recognition in humans relies on analysing shape, CNNs do not have such a shape-bias. We teased apart the type of features selected by the model by modifying the CIFAR-10 dataset so that, in addition to containing objects with shape, the images concurrently contained non-shape features, such as a noise-like mask. When trained on these modified set of images, the model did not show any bias towards selecting shapes as features. Instead it relied on whichever feature allowed it to perform the best prediction – even when this feature was a noise-like mask or a single predictive pixel amongst 50176 pixels. We also found that regularisation methods, such as batch normalisation or Dropout, did not change this behaviour and neither did past or concurrent experience with images from other datasets.
Tasks Object Recognition
Published 2019-05-01
URL https://openreview.net/forum?id=ByePUo05K7
PDF https://openreview.net/pdf?id=ByePUo05K7
PWC https://paperswithcode.com/paper/what-a-difference-a-pixel-makes-an-empirical
Repo
Framework

Joint extraction of entities and overlapping relations using position-attentive sequence labeling

Title Joint extraction of entities and overlapping relations using position-attentive sequence labeling
Authors Dai Dai, Xinyan Xiao, Yajuan Lyu, Shan Dou, Qiaoqiao She, Haifeng Wang
Abstract Joint entity and relation extraction is to detect entity and relation using a single model. In this paper, we present a novel unified joint extraction model which directly tags entity and relation labels according to a query word position p, i.e., detecting entity at p, and identifying entities at other positions that have relationship with the former. To this end, we first design a tagging scheme to generate n tag sequences for an n-word sentence. Then a position-attention mechanism is introduced to produce different sentence representations for every query position to model these n tag sequences. In this way, our method can simultaneously extract all entities and their type, as well as all overlapping relations. Experiment results show that our framework performances significantly better on extracting overlapping relations as well as detecting long-range relation, and thus we achieve state-of-the-art performance on two public datasets.
Tasks Joint Entity and Relation Extraction, Relation Extraction
Published 2019-07-17
URL https://www.aaai.org/ojs/index.php/AAAI/article/view/4591
PDF https://www.aaai.org/ojs/index.php/AAAI/article/view/4591/4469
PWC https://paperswithcode.com/paper/joint-extraction-of-entities-and-overlapping
Repo
Framework

Improving Robustness of Neural Machine Translation with Multi-task Learning

Title Improving Robustness of Neural Machine Translation with Multi-task Learning
Authors Shuyan Zhou, Xiangkai Zeng, Yingqi Zhou, Antonios Anastasopoulos, Graham Neubig
Abstract While neural machine translation (NMT) achieves remarkable performance on clean, in-domain text, performance is known to degrade drastically when facing text which is full of typos, grammatical errors and other varieties of noise. In this work, we propose a multi-task learning algorithm for transformer-based MT systems that is more resilient to this noise. We describe our submission to the WMT 2019 Robustness shared task based on this method. Our model achieves a BLEU score of 32.8 on the shared task French to English dataset, which is 7.1 BLEU points higher than the baseline vanilla transformer trained with clean text.
Tasks Machine Translation, Multi-Task Learning
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5368/
PDF https://www.aclweb.org/anthology/W19-5368
PWC https://paperswithcode.com/paper/improving-robustness-of-neural-machine
Repo
Framework

Nuclearity in RST and signals of coherence relations

Title Nuclearity in RST and signals of coherence relations
Authors Debopam Das
Abstract We investigate the relationship between the notion of nuclearity as proposed in Rhetorical Structure Theory (RST) and the signalling of coherence relations. RST relations are categorized as either mononuclear (comprising a nucleus and a satellite span) or multinuclear (comprising two or more nuclei spans). We examine how mononuclear relations (e.g., Antithesis, Condition) and multinuclear relations (e.g., Contrast, List) are indicated by relational signals, more particularly by discourse markers (e.g., because, however, if, therefore). We conduct a corpus study, examining the distribution of either type of relations in the RST Discourse Treebank (Carlson et al., 2002) and the distribution of discourse markers for those relations in the RST Signalling Corpus (Das et al., 2015). Our results show that discourse markers are used more often to signal multinuclear relations than mononuclear relations. The findings also suggest a complex relationship between the relation types and syntactic categories of discourse markers (subordinating and coordinating conjunctions).
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-2705/
PDF https://www.aclweb.org/anthology/W19-2705
PWC https://paperswithcode.com/paper/nuclearity-in-rst-and-signals-of-coherence
Repo
Framework

A Semantic Cover Approach for Topic Modeling

Title A Semantic Cover Approach for Topic Modeling
Authors Rajagopal Venkatesaramani, Doug Downey, Bradley Malin, Yevgeniy Vorobeychik
Abstract We introduce a novel topic modeling approach based on constructing a semantic set cover for clusters of similar documents. Specifically, our approach first clusters documents using their Tf-Idf representation, and then covers each cluster with a set of topic words based on semantic similarity, defined in terms of a word embedding. Computing a topic cover amounts to solving a minimum set cover problem. Our evaluation compares our topic modeling approach to Latent Dirichlet Allocation (LDA) on three metrics: 1) qualitative topic match, measured using evaluations by Amazon Mechanical Turk (MTurk) workers, 2) performance on classification tasks using each topic model as a sparse feature representation, and 3) topic coherence. We find that qualitative judgments significantly favor our approach, the method outperforms LDA on topic coherence, and is comparable to LDA on document classification tasks.
Tasks Document Classification, Semantic Similarity, Semantic Textual Similarity
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-1011/
PDF https://www.aclweb.org/anthology/S19-1011
PWC https://paperswithcode.com/paper/a-semantic-cover-approach-for-topic-modeling
Repo
Framework

R-STAN: Residual Spatial-Temporal Attention Network for Action Recognition

Title R-STAN: Residual Spatial-Temporal Attention Network for Action Recognition
Authors Quanle Liu, Xiangjiu Che, Mei Bie
Abstract Two-stream network architecture has the ability to capture temporal and spatial features from videos simultaneously and has achieved excellent performance on video action recognition tasks. However, there is a fair amount of redundant information in both temporal and spatial dimensions in videos, which increases the complexity of network learning. To solve this problem, we propose residual spatial-temporal attention network (R-STAN), a feed-forward convolutional neural network using residual learning and spatial-temporal attention mechanism for video action recognition, which makes the network focus more on discriminative temporal and spatial features. In our R-STAN, each stream is constructed by stacking residual spatial-temporal attention blocks (R-STAB), the spatial-temporal attention modules integrated in the residual blocks have the ability to generate attention-aware features along temporal and spatial dimensions, which largely reduce the redundant information. Together with the specific characteristic of residual learning, we are able to construct a very deep network for learning spatial-temporal information in videos. With the layers going deeper, the attention-aware features from the different R-STABs can change adaptively. We validate our R-STAN through a large number of experiments on UCF101 and HMDB51 datasets. Our experiments show that our proposed network combined with residual learning and spatial-temporal attention mechanism contributes substantially to the performance of video action recognition.
Tasks Action Recognition In Videos, Temporal Action Localization
Published 2019-06-19
URL https://doi.org/10.1109/ACCESS.2019.2923651
PDF https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=8740848
PWC https://paperswithcode.com/paper/r-stan-residual-spatial-temporal-attention
Repo
Framework

Event Detection with Multi-Order Graph Convolution and Aggregated Attention

Title Event Detection with Multi-Order Graph Convolution and Aggregated Attention
Authors Haoran Yan, Xiaolong Jin, Xiangbin Meng, Jiafeng Guo, Xueqi Cheng
Abstract Syntactic relations are broadly used in many NLP tasks. For event detection, syntactic relation representations based on dependency tree can better capture the interrelations between candidate trigger words and related entities than sentence representations. But, existing studies only use first-order syntactic relations (i.e., the arcs) in dependency trees to identify trigger words. For this reason, this paper proposes a new method for event detection, which uses a dependency tree based graph convolution network with aggregative attention to explicitly model and aggregate multi-order syntactic representations in sentences. Experimental comparison with state-of-the-art baselines shows the superiority of the proposed method.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1582/
PDF https://www.aclweb.org/anthology/D19-1582
PWC https://paperswithcode.com/paper/event-detection-with-multi-order-graph
Repo
Framework

MLGCN: Multi-Laplacian Graph Convolutional Networks for Human Action Recognition

Title MLGCN: Multi-Laplacian Graph Convolutional Networks for Human Action Recognition
Authors Ahmed Mazari, Hichem Sahbi
Abstract Convolutional neural networks are nowadays witnessing a major success in different pattern recognition problems. These learning models were basically designed to handle vectorial data such as images but their extension to non-vectorial and semi-structured data (namely graphs with variable sizes, topology, etc.) remains a major challenge, though a few interesting solutions are currently emerging. In this paper, we introduce MLGCN; a novel spectral Multi-Laplacian Graph Convolutional Network. The main ontribution of this method resides in a new design principle that learns graph-laplacians as convex combinations of other elementary laplacians – each one dedicated to a particular topology of the input graphs. We also introduce a novel pooling operator, on graphs, that proceeds in two steps: context-dependent node expansion is achieved, followed by a global average pooling; the strength of this two-step process resides in its ability to preserve the discrimination power of nodes while achieving permutation invariance. Experiments conducted on SBU and UCF-101 datasets, show the validity of our method for the challenging task of action recognition. Supplementary : https://bit.ly/2ku2lYv
Tasks Action Recognition In Videos, Skeleton Based Action Recognition, Temporal Action Localization
Published 2019-09-11
URL https://bmvc2019.org/wp-content/uploads/papers/1103-paper.pdf
PDF https://bmvc2019.org/wp-content/uploads/papers/1103-paper.pdf
PWC https://paperswithcode.com/paper/mlgcn-multi-laplacian-graph-convolutional
Repo
Framework

Cross-lingual Transfer Learning for Japanese Named Entity Recognition

Title Cross-lingual Transfer Learning for Japanese Named Entity Recognition
Authors Andrew Johnson, Penny Karanasou, Judith Gaspers, Dietrich Klakow
Abstract This work explores cross-lingual transfer learning (TL) for named entity recognition, focusing on bootstrapping Japanese from English. A deep neural network model is adopted and the best combination of weights to transfer is extensively investigated. Moreover, a novel approach is presented that overcomes linguistic differences between this language pair by romanizing a portion of the Japanese input. Experiments are conducted on external datasets, as well as internal large-scale real-world ones. Gains with TL are achieved for all evaluated cases. Finally, the influence on TL of the target dataset size and of the target tagset distribution is further investigated.
Tasks Cross-Lingual Transfer, Named Entity Recognition, Transfer Learning
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-2023/
PDF https://www.aclweb.org/anthology/N19-2023
PWC https://paperswithcode.com/paper/cross-lingual-transfer-learning-for-japanese
Repo
Framework

Synthesizing Environment-Aware Activities via Activity Sketches

Title Synthesizing Environment-Aware Activities via Activity Sketches
Authors Yuan-Hong Liao, Xavier Puig, Marko Boben, Antonio Torralba, Sanja Fidler
Abstract In order to learn to perform activities from demonstrations or descriptions, agents need to distill what the essence of the given activity is, and how it can be adapted to new environments. In this work, we address the problem: environment-aware program generation. Given a visual demonstration or a description of an activity, we generate program sketches representing the essential instructions and propose a model to flesh these into full programs representing the actions needed to perform the activity under the presented environmental constraints. To this end, we build upon VirtualHome, to create a new dataset VirtualHome-Env, where we collect program sketches to represent activities and match programs with environments that can afford them. Furthermore, we construct a knowledge base to sample realistic environments and another knowledge base to seek out the programs under the sampled environments. Finally, we propose RNN-ResActGraph, a network that generates a program from a given sketch and an environment graph and tracks the changes in the environment induced by the program.
Tasks
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Liao_Synthesizing_Environment-Aware_Activities_via_Activity_Sketches_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Liao_Synthesizing_Environment-Aware_Activities_via_Activity_Sketches_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/synthesizing-environment-aware-activities-via
Repo
Framework

Learning Actionable Representations with Goal Conditioned Policies

Title Learning Actionable Representations with Goal Conditioned Policies
Authors Dibya Ghosh, Abhishek Gupta, Sergey Levine
Abstract Representation learning is a central challenge across a range of machine learning areas. In reinforcement learning, effective and functional representations have the potential to tremendously accelerate learning progress and solve more challenging problems. Most prior work on representation learning has focused on generative approaches, learning representations that capture all the underlying factors of variation in the observation space in a more disentangled or well-ordered manner. In this paper, we instead aim to learn functionally salient representations: representations that are not necessarily complete in terms of capturing all factors of variation in the observation space, but rather aim to capture those factors of variation that are important for decision making – that are “actionable”. These representations are aware of the dynamics of the environment, and capture only the elements of the observation that are necessary for decision making rather than all factors of variation, eliminating the need for explicit reconstruction. We show how these learned representations can be useful to improve exploration for sparse reward problems, to enable long horizon hierarchical reinforcement learning, and as a state representation for learning policies for downstream tasks. We evaluate our method on a number of simulated environments, and compare it to prior methods for representation learning, exploration, and hierarchical reinforcement learning.
Tasks Decision Making, Hierarchical Reinforcement Learning, Representation Learning
Published 2019-05-01
URL https://openreview.net/forum?id=Hye9lnCct7
PDF https://openreview.net/pdf?id=Hye9lnCct7
PWC https://paperswithcode.com/paper/learning-actionable-representations-with-goal-1
Repo
Framework
comments powered by Disqus