Paper Group AWR 126
Natural Language Does Not Emerge ‘Naturally’ in Multi-Agent Dialog. Variational Approaches for Auto-Encoding Generative Adversarial Networks. Transform-Invariant Non-Parametric Clustering of Covariance Matrices and its Application to Unsupervised Joint Segmentation and Action Discovery. Safe Policy Improvement with Baseline Bootstrapping. Guided De …
Natural Language Does Not Emerge ‘Naturally’ in Multi-Agent Dialog
Title | Natural Language Does Not Emerge ‘Naturally’ in Multi-Agent Dialog |
Authors | Satwik Kottur, José M. F. Moura, Stefan Lee, Dhruv Batra |
Abstract | A number of recent works have proposed techniques for end-to-end learning of communication protocols among cooperative multi-agent populations, and have simultaneously found the emergence of grounded human-interpretable language in the protocols developed by the agents, all learned without any human supervision! In this paper, using a Task and Tell reference game between two agents as a testbed, we present a sequence of ‘negative’ results culminating in a ‘positive’ one – showing that while most agent-invented languages are effective (i.e. achieve near-perfect task rewards), they are decidedly not interpretable or compositional. In essence, we find that natural language does not emerge ‘naturally’, despite the semblance of ease of natural-language-emergence that one may gather from recent literature. We discuss how it is possible to coax the invented languages to become more and more human-like and compositional by increasing restrictions on how two agents may communicate. |
Tasks | |
Published | 2017-06-26 |
URL | http://arxiv.org/abs/1706.08502v3 |
http://arxiv.org/pdf/1706.08502v3.pdf | |
PWC | https://paperswithcode.com/paper/natural-language-does-not-emerge-naturally-in |
Repo | https://github.com/kdexd/lang-emerge-parlai |
Framework | pytorch |
Variational Approaches for Auto-Encoding Generative Adversarial Networks
Title | Variational Approaches for Auto-Encoding Generative Adversarial Networks |
Authors | Mihaela Rosca, Balaji Lakshminarayanan, David Warde-Farley, Shakir Mohamed |
Abstract | Auto-encoding generative adversarial networks (GANs) combine the standard GAN algorithm, which discriminates between real and model-generated data, with a reconstruction loss given by an auto-encoder. Such models aim to prevent mode collapse in the learned generative model by ensuring that it is grounded in all the available training data. In this paper, we develop a principle upon which auto-encoders can be combined with generative adversarial networks by exploiting the hierarchical structure of the generative model. The underlying principle shows that variational inference can be used a basic tool for learning, but with the in- tractable likelihood replaced by a synthetic likelihood, and the unknown posterior distribution replaced by an implicit distribution; both synthetic likelihoods and implicit posterior distributions can be learned using discriminators. This allows us to develop a natural fusion of variational auto-encoders and generative adversarial networks, combining the best of both these methods. We describe a unified objective for optimization, discuss the constraints needed to guide learning, connect to the wide range of existing work, and use a battery of tests to systematically and quantitatively assess the performance of our method. |
Tasks | |
Published | 2017-06-15 |
URL | http://arxiv.org/abs/1706.04987v2 |
http://arxiv.org/pdf/1706.04987v2.pdf | |
PWC | https://paperswithcode.com/paper/variational-approaches-for-auto-encoding |
Repo | https://github.com/PrateekMunjal/Alpha_GAN |
Framework | tf |
Transform-Invariant Non-Parametric Clustering of Covariance Matrices and its Application to Unsupervised Joint Segmentation and Action Discovery
Title | Transform-Invariant Non-Parametric Clustering of Covariance Matrices and its Application to Unsupervised Joint Segmentation and Action Discovery |
Authors | Nadia Figueroa, Aude Billard |
Abstract | In this work, we tackle the problem of transform-invariant unsupervised learning in the space of Covariance matrices and applications thereof. We begin by introducing the Spectral Polytope Covariance Matrix (SPCM) Similarity function; a similarity function for Covariance matrices, invariant to any type of transformation. We then derive the SPCM-CRP mixture model, a transform-invariant non-parametric clustering approach for Covariance matrices that leverages the proposed similarity function, spectral embedding and the distance-dependent Chinese Restaurant Process (dd-CRP) (Blei and Frazier, 2011). The scalability and applicability of these two contributions is extensively validated on real-world Covariance matrix datasets from diverse research fields. Finally, we couple the SPCM-CRP mixture model with the Bayesian non-parametric Indian Buffet Process (IBP) - Hidden Markov Model (HMM) (Fox et al., 2009), to jointly segment and discover transform-invariant action primitives from complex sequential data. Resulting in a topic-modeling inspired hierarchical model for unsupervised time-series data analysis which we call ICSC-HMM (IBP Coupled SPCM-CRP Hidden Markov Model). The ICSC-HMM is validated on kinesthetic demonstrations of uni-manual and bi-manual cooking tasks; achieving unsupervised human-level decomposition of complex sequential tasks. |
Tasks | Time Series |
Published | 2017-10-27 |
URL | http://arxiv.org/abs/1710.10060v1 |
http://arxiv.org/pdf/1710.10060v1.pdf | |
PWC | https://paperswithcode.com/paper/transform-invariant-non-parametric-clustering |
Repo | https://github.com/nbfigueroa/SPCM-CRP |
Framework | none |
Safe Policy Improvement with Baseline Bootstrapping
Title | Safe Policy Improvement with Baseline Bootstrapping |
Authors | Romain Laroche, Paul Trichelair, Rémi Tachet des Combes |
Abstract | This paper considers Safe Policy Improvement (SPI) in Batch Reinforcement Learning (Batch RL): from a fixed dataset and without direct access to the true environment, train a policy that is guaranteed to perform at least as well as the baseline policy used to collect the data. Our approach, called SPI with Baseline Bootstrapping (SPIBB), is inspired by the knows-what-it-knows paradigm: it bootstraps the trained policy with the baseline when the uncertainty is high. Our first algorithm, $\Pi_b$-SPIBB, comes with SPI theoretical guarantees. We also implement a variant, $\Pi_{\leq b}$-SPIBB, that is even more efficient in practice. We apply our algorithms to a motivational stochastic gridworld domain and further demonstrate on randomly generated MDPs the superiority of SPIBB with respect to existing algorithms, not only in safety but also in mean performance. Finally, we implement a model-free version of SPIBB and show its benefits on a navigation task with deep RL implementation called SPIBB-DQN, which is, to the best of our knowledge, the first RL algorithm relying on a neural network representation able to train efficiently and reliably from batch data, without any interaction with the environment. |
Tasks | |
Published | 2017-12-19 |
URL | https://arxiv.org/abs/1712.06924v5 |
https://arxiv.org/pdf/1712.06924v5.pdf | |
PWC | https://paperswithcode.com/paper/safe-policy-improvement-with-baseline |
Repo | https://github.com/RomainLaroche/SPIBB |
Framework | none |
Guided Deep Reinforcement Learning for Swarm Systems
Title | Guided Deep Reinforcement Learning for Swarm Systems |
Authors | Maximilian Hüttenrauch, Adrian Šošić, Gerhard Neumann |
Abstract | In this paper, we investigate how to learn to control a group of cooperative agents with limited sensing capabilities such as robot swarms. The agents have only very basic sensor capabilities, yet in a group they can accomplish sophisticated tasks, such as distributed assembly or search and rescue tasks. Learning a policy for a group of agents is difficult due to distributed partial observability of the state. Here, we follow a guided approach where a critic has central access to the global state during learning, which simplifies the policy evaluation problem from a reinforcement learning point of view. For example, we can get the positions of all robots of the swarm using a camera image of a scene. This camera image is only available to the critic and not to the control policies of the robots. We follow an actor-critic approach, where the actors base their decisions only on locally sensed information. In contrast, the critic is learned based on the true global state. Our algorithm uses deep reinforcement learning to approximate both the Q-function and the policy. The performance of the algorithm is evaluated on two tasks with simple simulated 2D agents: 1) finding and maintaining a certain distance to each others and 2) locating a target. |
Tasks | |
Published | 2017-09-18 |
URL | http://arxiv.org/abs/1709.06011v1 |
http://arxiv.org/pdf/1709.06011v1.pdf | |
PWC | https://paperswithcode.com/paper/guided-deep-reinforcement-learning-for-swarm |
Repo | https://github.com/nsrishankar/rl_swarm_papers |
Framework | none |
RaspiReader: Open Source Fingerprint Reader
Title | RaspiReader: Open Source Fingerprint Reader |
Authors | Joshua J. Engelsma, Kai Cao, Anil K. Jain |
Abstract | We open source an easy to assemble, spoof resistant, high resolution, optical fingerprint reader, called RaspiReader, using ubiquitous components. By using our open source STL files and software, RaspiReader can be built in under one hour for only US $175. As such, RaspiReader provides the fingerprint research community a seamless and simple method for quickly prototyping new ideas involving fingerprint reader hardware. In particular, we posit that this open source fingerprint reader will facilitate the exploration of novel fingerprint spoof detection techniques involving both hardware and software. We demonstrate one such spoof detection technique by specially customizing RaspiReader with two cameras for fingerprint image acquisition. One camera provides high contrast, frustrated total internal reflection (FTIR) fingerprint images, and the other outputs direct images of the finger in contact with the platen. Using both of these image streams, we extract complementary information which, when fused together and used for spoof detection, results in marked performance improvement over previous methods relying only on grayscale FTIR images provided by COTS optical readers. Finally, fingerprint matching experiments between images acquired from the FTIR output of RaspiReader and images acquired from a COTS reader verify the interoperability of the RaspiReader with existing COTS optical readers. |
Tasks | |
Published | 2017-12-26 |
URL | http://arxiv.org/abs/1712.09392v1 |
http://arxiv.org/pdf/1712.09392v1.pdf | |
PWC | https://paperswithcode.com/paper/raspireader-open-source-fingerprint-reader |
Repo | https://github.com/engelsjo/RaspiReader |
Framework | none |
Functional Map of the World
Title | Functional Map of the World |
Authors | Gordon Christie, Neil Fendley, James Wilson, Ryan Mukherjee |
Abstract | We present a new dataset, Functional Map of the World (fMoW), which aims to inspire the development of machine learning models capable of predicting the functional purpose of buildings and land use from temporal sequences of satellite images and a rich set of metadata features. The metadata provided with each image enables reasoning about location, time, sun angles, physical sizes, and other features when making predictions about objects in the image. Our dataset consists of over 1 million images from over 200 countries. For each image, we provide at least one bounding box annotation containing one of 63 categories, including a “false detection” category. We present an analysis of the dataset along with baseline approaches that reason about metadata and temporal views. Our data, code, and pretrained models have been made publicly available. |
Tasks | |
Published | 2017-11-21 |
URL | http://arxiv.org/abs/1711.07846v3 |
http://arxiv.org/pdf/1711.07846v3.pdf | |
PWC | https://paperswithcode.com/paper/functional-map-of-the-world |
Repo | https://github.com/fmow/baseline |
Framework | tf |
APE-GAN: Adversarial Perturbation Elimination with GAN
Title | APE-GAN: Adversarial Perturbation Elimination with GAN |
Authors | Shiwei Shen, Guoqing Jin, Ke Gao, Yongdong Zhang |
Abstract | Although neural networks could achieve state-of-the-art performance while recongnizing images, they often suffer a tremendous defeat from adversarial examples–inputs generated by utilizing imperceptible but intentional perturbation to clean samples from the datasets. How to defense against adversarial examples is an important problem which is well worth researching. So far, very few methods have provided a significant defense to adversarial examples. In this paper, a novel idea is proposed and an effective framework based Generative Adversarial Nets named APE-GAN is implemented to defense against the adversarial examples. The experimental results on three benchmark datasets including MNIST, CIFAR10 and ImageNet indicate that APE-GAN is effective to resist adversarial examples generated from five attacks. |
Tasks | |
Published | 2017-07-18 |
URL | http://arxiv.org/abs/1707.05474v3 |
http://arxiv.org/pdf/1707.05474v3.pdf | |
PWC | https://paperswithcode.com/paper/ape-gan-adversarial-perturbation-elimination |
Repo | https://github.com/carlini/APE-GAN |
Framework | tf |
Behavior Trees in Robotics and AI: An Introduction
Title | Behavior Trees in Robotics and AI: An Introduction |
Authors | Michele Colledanchise, Petter Ögren |
Abstract | A Behavior Tree (BT) is a way to structure the switching between different tasks in an autonomous agent, such as a robot or a virtual entity in a computer game. BTs are a very efficient way of creating complex systems that are both modular and reactive. These properties are crucial in many applications, which has led to the spread of BT from computer game programming to many branches of AI and Robotics. In this book, we will first give an introduction to BTs, then we describe how BTs relate to, and in many cases generalize, earlier switching structures. These ideas are then used as a foundation for a set of efficient and easy to use design principles. Properties such as safety, robustness, and efficiency are important for an autonomous system, and we describe a set of tools for formally analyzing these using a state space description of BTs. With the new analysis tools, we can formalize the descriptions of how BTs generalize earlier approaches. We also show the use of BTs in automated planning and machine learning. Finally, we describe an extended set of tools to capture the behavior of Stochastic BTs, where the outcomes of actions are described by probabilities. These tools enable the computation of both success probabilities and time to completion. |
Tasks | |
Published | 2017-08-31 |
URL | http://arxiv.org/abs/1709.00084v3 |
http://arxiv.org/pdf/1709.00084v3.pdf | |
PWC | https://paperswithcode.com/paper/behavior-trees-in-robotics-and-ai-an |
Repo | https://github.com/miccol/ROS-Behavior-Tree |
Framework | none |
Automatically Generating Commit Messages from Diffs using Neural Machine Translation
Title | Automatically Generating Commit Messages from Diffs using Neural Machine Translation |
Authors | Siyuan Jiang, Ameer Armaly, Collin McMillan |
Abstract | Commit messages are a valuable resource in comprehension of software evolution, since they provide a record of changes such as feature additions and bug repairs. Unfortunately, programmers often neglect to write good commit messages. Different techniques have been proposed to help programmers by automatically writing these messages. These techniques are effective at describing what changed, but are often verbose and lack context for understanding the rationale behind a change. In contrast, humans write messages that are short and summarize the high level rationale. In this paper, we adapt Neural Machine Translation (NMT) to automatically “translate” diffs into commit messages. We trained an NMT algorithm using a corpus of diffs and human-written commit messages from the top 1k Github projects. We designed a filter to help ensure that we only trained the algorithm on higher-quality commit messages. Our evaluation uncovered a pattern in which the messages we generate tend to be either very high or very low quality. Therefore, we created a quality-assurance filter to detect cases in which we are unable to produce good messages, and return a warning instead. |
Tasks | Machine Translation |
Published | 2017-08-30 |
URL | http://arxiv.org/abs/1708.09492v1 |
http://arxiv.org/pdf/1708.09492v1.pdf | |
PWC | https://paperswithcode.com/paper/automatically-generating-commit-messages-from |
Repo | https://github.com/vladislavneon/autogenerating-commit-messages |
Framework | none |
STN-OCR: A single Neural Network for Text Detection and Text Recognition
Title | STN-OCR: A single Neural Network for Text Detection and Text Recognition |
Authors | Christian Bartz, Haojin Yang, Christoph Meinel |
Abstract | Detecting and recognizing text in natural scene images is a challenging, yet not completely solved task. In re- cent years several new systems that try to solve at least one of the two sub-tasks (text detection and text recognition) have been proposed. In this paper we present STN-OCR, a step towards semi-supervised neural networks for scene text recognition, that can be optimized end-to-end. In contrast to most existing works that consist of multiple deep neural networks and several pre-processing steps we propose to use a single deep neural network that learns to detect and recognize text from natural images in a semi-supervised way. STN-OCR is a network that integrates and jointly learns a spatial transformer network, that can learn to detect text regions in an image, and a text recognition network that takes the identified text regions and recognizes their textual content. We investigate how our model behaves on a range of different tasks (detection and recognition of characters, and lines of text). Experimental results on public benchmark datasets show the ability of our model to handle a variety of different tasks, without substantial changes in its overall network structure. |
Tasks | Optical Character Recognition, Scene Text Detection, Scene Text Recognition |
Published | 2017-07-27 |
URL | http://arxiv.org/abs/1707.08831v1 |
http://arxiv.org/pdf/1707.08831v1.pdf | |
PWC | https://paperswithcode.com/paper/stn-ocr-a-single-neural-network-for-text |
Repo | https://github.com/Narp99/FRI_2018 |
Framework | tf |
Chinese Typography Transfer
Title | Chinese Typography Transfer |
Authors | Jie Chang, Yujun Gu |
Abstract | In this paper, we propose a new network architecture for Chinese typography transformation based on deep learning. The architecture consists of two sub-networks: (1)a fully convolutional network(FCN) aiming at transferring specified typography style to another in condition of preserving structure information; (2)an adversarial network aiming at generating more realistic strokes in some details. Unlike models proposed before 2012 relying on the complex segmentation of Chinese components or strokes, our model treats every Chinese character as an inseparable image, so pre-processing or post-preprocessing are abandoned. Besides, our model adopts end-to-end training without pre-trained used in other deep models. The experiments demonstrates that our model can synthesize realistic-looking target typography from any source typography both on printed style and handwriting style. |
Tasks | |
Published | 2017-07-16 |
URL | http://arxiv.org/abs/1707.04904v2 |
http://arxiv.org/pdf/1707.04904v2.pdf | |
PWC | https://paperswithcode.com/paper/chinese-typography-transfer |
Repo | https://github.com/hanfeisun/Unsupervised-Typography-Transfer |
Framework | tf |
Fast Spectral Clustering Using Autoencoders and Landmarks
Title | Fast Spectral Clustering Using Autoencoders and Landmarks |
Authors | Ershad Banijamali, Ali Ghodsi |
Abstract | In this paper, we introduce an algorithm for performing spectral clustering efficiently. Spectral clustering is a powerful clustering algorithm that suffers from high computational complexity, due to eigen decomposition. In this work, we first build the adjacency matrix of the corresponding graph of the dataset. To build this matrix, we only consider a limited number of points, called landmarks, and compute the similarity of all data points with the landmarks. Then, we present a definition of the Laplacian matrix of the graph that enable us to perform eigen decomposition efficiently, using a deep autoencoder. The overall complexity of the algorithm for eigen decomposition is $O(np)$, where $n$ is the number of data points and $p$ is the number of landmarks. At last, we evaluate the performance of the algorithm in different experiments. |
Tasks | |
Published | 2017-04-07 |
URL | http://arxiv.org/abs/1704.02345v1 |
http://arxiv.org/pdf/1704.02345v1.pdf | |
PWC | https://paperswithcode.com/paper/fast-spectral-clustering-using-autoencoders |
Repo | https://github.com/chenjs12/ML |
Framework | none |
Dual Rectified Linear Units (DReLUs): A Replacement for Tanh Activation Functions in Quasi-Recurrent Neural Networks
Title | Dual Rectified Linear Units (DReLUs): A Replacement for Tanh Activation Functions in Quasi-Recurrent Neural Networks |
Authors | Fréderic Godin, Jonas Degrave, Joni Dambre, Wesley De Neve |
Abstract | In this paper, we introduce a novel type of Rectified Linear Unit (ReLU), called a Dual Rectified Linear Unit (DReLU). A DReLU, which comes with an unbounded positive and negative image, can be used as a drop-in replacement for a tanh activation function in the recurrent step of Quasi-Recurrent Neural Networks (QRNNs) (Bradbury et al. (2017)). Similar to ReLUs, DReLUs are less prone to the vanishing gradient problem, they are noise robust, and they induce sparse activations. We independently reproduce the QRNN experiments of Bradbury et al. (2017) and compare our DReLU-based QRNNs with the original tanh-based QRNNs and Long Short-Term Memory networks (LSTMs) on sentiment classification and word-level language modeling. Additionally, we evaluate on character-level language modeling, showing that we are able to stack up to eight QRNN layers with DReLUs, thus making it possible to improve the current state-of-the-art in character-level language modeling over shallow architectures based on LSTMs. |
Tasks | Language Modelling, Sentiment Analysis |
Published | 2017-07-25 |
URL | http://arxiv.org/abs/1707.08214v2 |
http://arxiv.org/pdf/1707.08214v2.pdf | |
PWC | https://paperswithcode.com/paper/dual-rectified-linear-units-drelus-a |
Repo | https://github.com/FredericGodin/QuasiRNN-DReLU |
Framework | none |
Open Set Logo Detection and Retrieval
Title | Open Set Logo Detection and Retrieval |
Authors | Andras Tüzkö, Christian Herrmann, Daniel Manger, Jürgen Beyerer |
Abstract | Current logo retrieval research focuses on closed set scenarios. We argue that the logo domain is too large for this strategy and requires an open set approach. To foster research in this direction, a large-scale logo dataset, called Logos in the Wild, is collected and released to the public. A typical open set logo retrieval application is, for example, assessing the effectiveness of advertisement in sports event broadcasts. Given a query sample in shape of a logo image, the task is to find all further occurrences of this logo in a set of images or videos. Currently, common logo retrieval approaches are unsuitable for this task because of their closed world assumption. Thus, an open set logo retrieval method is proposed in this work which allows searching for previously unseen logos by a single query sample. A two stage concept with separate logo detection and comparison is proposed where both modules are based on task specific CNNs. If trained with the Logos in the Wild data, significant performance improvements are observed, especially compared with state-of-the-art closed set approaches. |
Tasks | |
Published | 2017-10-30 |
URL | http://arxiv.org/abs/1710.10891v1 |
http://arxiv.org/pdf/1710.10891v1.pdf | |
PWC | https://paperswithcode.com/paper/open-set-logo-detection-and-retrieval |
Repo | https://github.com/ilmonteux/logohunter |
Framework | tf |