Paper Group AWR 152
Robust Adversarial Reinforcement Learning. Unsupervised Anomaly Detection with Generative Adversarial Networks to Guide Marker Discovery. Breast density classification with deep convolutional neural networks. Zero-Shot Relation Extraction via Reading Comprehension. DyNet: The Dynamic Neural Network Toolkit. An Optimization Approach to Learning Fall …
Robust Adversarial Reinforcement Learning
Title | Robust Adversarial Reinforcement Learning |
Authors | Lerrel Pinto, James Davidson, Rahul Sukthankar, Abhinav Gupta |
Abstract | Deep neural networks coupled with fast simulation and improved computation have led to recent successes in the field of reinforcement learning (RL). However, most current RL-based approaches fail to generalize since: (a) the gap between simulation and real world is so large that policy-learning approaches fail to transfer; (b) even if policy learning is done in real world, the data scarcity leads to failed generalization from training to test scenarios (e.g., due to different friction or object masses). Inspired from H-infinity control methods, we note that both modeling errors and differences in training and test scenarios can be viewed as extra forces/disturbances in the system. This paper proposes the idea of robust adversarial reinforcement learning (RARL), where we train an agent to operate in the presence of a destabilizing adversary that applies disturbance forces to the system. The jointly trained adversary is reinforced – that is, it learns an optimal destabilization policy. We formulate the policy learning as a zero-sum, minimax objective function. Extensive experiments in multiple environments (InvertedPendulum, HalfCheetah, Swimmer, Hopper and Walker2d) conclusively demonstrate that our method (a) improves training stability; (b) is robust to differences in training/test conditions; and c) outperform the baseline even in the absence of the adversary. |
Tasks | |
Published | 2017-03-08 |
URL | http://arxiv.org/abs/1703.02702v1 |
http://arxiv.org/pdf/1703.02702v1.pdf | |
PWC | https://paperswithcode.com/paper/robust-adversarial-reinforcement-learning |
Repo | https://github.com/davidsonic/robust-grasp |
Framework | tf |
Unsupervised Anomaly Detection with Generative Adversarial Networks to Guide Marker Discovery
Title | Unsupervised Anomaly Detection with Generative Adversarial Networks to Guide Marker Discovery |
Authors | Thomas Schlegl, Philipp Seeböck, Sebastian M. Waldstein, Ursula Schmidt-Erfurth, Georg Langs |
Abstract | Obtaining models that capture imaging markers relevant for disease progression and treatment monitoring is challenging. Models are typically based on large amounts of data with annotated examples of known markers aiming at automating detection. High annotation effort and the limitation to a vocabulary of known markers limit the power of such approaches. Here, we perform unsupervised learning to identify anomalies in imaging data as candidates for markers. We propose AnoGAN, a deep convolutional generative adversarial network to learn a manifold of normal anatomical variability, accompanying a novel anomaly scoring scheme based on the mapping from image space to a latent space. Applied to new data, the model labels anomalies, and scores image patches indicating their fit into the learned distribution. Results on optical coherence tomography images of the retina demonstrate that the approach correctly identifies anomalous images, such as images containing retinal fluid or hyperreflective foci. |
Tasks | Anomaly Detection, Unsupervised Anomaly Detection |
Published | 2017-03-17 |
URL | http://arxiv.org/abs/1703.05921v1 |
http://arxiv.org/pdf/1703.05921v1.pdf | |
PWC | https://paperswithcode.com/paper/unsupervised-anomaly-detection-with |
Repo | https://github.com/seungjunlee96/DeepLearning-for-Medical-Imaging |
Framework | none |
Breast density classification with deep convolutional neural networks
Title | Breast density classification with deep convolutional neural networks |
Authors | Nan Wu, Krzysztof J. Geras, Yiqiu Shen, Jingyi Su, S. Gene Kim, Eric Kim, Stacey Wolfson, Linda Moy, Kyunghyun Cho |
Abstract | Breast density classification is an essential part of breast cancer screening. Although a lot of prior work considered this problem as a task for learning algorithms, to our knowledge, all of them used small and not clinically realistic data both for training and evaluation of their models. In this work, we explore the limits of this task with a data set coming from over 200,000 breast cancer screening exams. We use this data to train and evaluate a strong convolutional neural network classifier. In a reader study, we find that our model can perform this task comparably to a human expert. |
Tasks | breast density classification |
Published | 2017-11-10 |
URL | http://arxiv.org/abs/1711.03674v1 |
http://arxiv.org/pdf/1711.03674v1.pdf | |
PWC | https://paperswithcode.com/paper/breast-density-classification-with-deep |
Repo | https://github.com/nyukat/breast_density_classifier |
Framework | pytorch |
Zero-Shot Relation Extraction via Reading Comprehension
Title | Zero-Shot Relation Extraction via Reading Comprehension |
Authors | Omer Levy, Minjoon Seo, Eunsol Choi, Luke Zettlemoyer |
Abstract | We show that relation extraction can be reduced to answering simple reading comprehension questions, by associating one or more natural-language questions with each relation slot. This reduction has several advantages: we can (1) learn relation-extraction models by extending recent neural reading-comprehension techniques, (2) build very large training sets for those models by combining relation-specific crowd-sourced questions with distant supervision, and even (3) do zero-shot learning by extracting new relation types that are only specified at test-time, for which we have no labeled training examples. Experiments on a Wikipedia slot-filling task demonstrate that the approach can generalize to new questions for known relation types with high accuracy, and that zero-shot generalization to unseen relation types is possible, at lower accuracy levels, setting the bar for future work on this task. |
Tasks | Reading Comprehension, Relation Extraction, Slot Filling, Zero-Shot Learning |
Published | 2017-06-13 |
URL | http://arxiv.org/abs/1706.04115v1 |
http://arxiv.org/pdf/1706.04115v1.pdf | |
PWC | https://paperswithcode.com/paper/zero-shot-relation-extraction-via-reading |
Repo | https://github.com/zhuzhicai/SQuAD2.0-Baseline-Test-with-BiDAF-No-Answer |
Framework | tf |
DyNet: The Dynamic Neural Network Toolkit
Title | DyNet: The Dynamic Neural Network Toolkit |
Authors | Graham Neubig, Chris Dyer, Yoav Goldberg, Austin Matthews, Waleed Ammar, Antonios Anastasopoulos, Miguel Ballesteros, David Chiang, Daniel Clothiaux, Trevor Cohn, Kevin Duh, Manaal Faruqui, Cynthia Gan, Dan Garrette, Yangfeng Ji, Lingpeng Kong, Adhiguna Kuncoro, Gaurav Kumar, Chaitanya Malaviya, Paul Michel, Yusuke Oda, Matthew Richardson, Naomi Saphra, Swabha Swayamdipta, Pengcheng Yin |
Abstract | We describe DyNet, a toolkit for implementing neural network models based on dynamic declaration of network structure. In the static declaration strategy that is used in toolkits like Theano, CNTK, and TensorFlow, the user first defines a computation graph (a symbolic representation of the computation), and then examples are fed into an engine that executes this computation and computes its derivatives. In DyNet’s dynamic declaration strategy, computation graph construction is mostly transparent, being implicitly constructed by executing procedural code that computes the network outputs, and the user is free to use different network structures for each input. Dynamic declaration thus facilitates the implementation of more complicated network architectures, and DyNet is specifically designed to allow users to implement their models in a way that is idiomatic in their preferred programming language (C++ or Python). One challenge with dynamic declaration is that because the symbolic computation graph is defined anew for every training example, its construction must have low overhead. To achieve this, DyNet has an optimized C++ backend and lightweight graph representation. Experiments show that DyNet’s speeds are faster than or comparable with static declaration toolkits, and significantly faster than Chainer, another dynamic declaration toolkit. DyNet is released open-source under the Apache 2.0 license and available at http://github.com/clab/dynet. |
Tasks | graph construction |
Published | 2017-01-15 |
URL | http://arxiv.org/abs/1701.03980v1 |
http://arxiv.org/pdf/1701.03980v1.pdf | |
PWC | https://paperswithcode.com/paper/dynet-the-dynamic-neural-network-toolkit |
Repo | https://github.com/hanyaqian/dynet |
Framework | none |
An Optimization Approach to Learning Falling Rule Lists
Title | An Optimization Approach to Learning Falling Rule Lists |
Authors | Chaofan Chen, Cynthia Rudin |
Abstract | A falling rule list is a probabilistic decision list for binary classification, consisting of a series of if-then rules with antecedents in the if clauses and probabilities of the desired outcome (“1”) in the then clauses. Just as in a regular decision list, the order of rules in a falling rule list is important – each example is classified by the first rule whose antecedent it satisfies. Unlike a regular decision list, a falling rule list requires the probabilities of the desired outcome (“1”) to be monotonically decreasing down the list. We propose an optimization approach to learning falling rule lists and “softly” falling rule lists, along with Monte-Carlo search algorithms that use bounds on the optimal solution to prune the search space. |
Tasks | |
Published | 2017-10-06 |
URL | http://arxiv.org/abs/1710.02572v3 |
http://arxiv.org/pdf/1710.02572v3.pdf | |
PWC | https://paperswithcode.com/paper/an-optimization-approach-to-learning-falling |
Repo | https://github.com/cfchen-duke/FRLOptimization |
Framework | none |
CNN features are also great at unsupervised classification
Title | CNN features are also great at unsupervised classification |
Authors | Joris Guérin, Olivier Gibaru, Stéphane Thiery, Eric Nyiri |
Abstract | This paper aims at providing insight on the transferability of deep CNN features to unsupervised problems. We study the impact of different pretrained CNN feature extractors on the problem of image set clustering for object classification as well as fine-grained classification. We propose a rather straightforward pipeline combining deep-feature extraction using a CNN pretrained on ImageNet and a classic clustering algorithm to classify sets of images. This approach is compared to state-of-the-art algorithms in image-clustering and provides better results. These results strengthen the belief that supervised training of deep CNN on large datasets, with a large variability of classes, extracts better features than most carefully designed engineering approaches, even for unsupervised tasks. We also validate our approach on a robotic application, consisting in sorting and storing objects smartly based on clustering. |
Tasks | Image Clustering, Object Classification |
Published | 2017-07-06 |
URL | http://arxiv.org/abs/1707.01700v2 |
http://arxiv.org/pdf/1707.01700v2.pdf | |
PWC | https://paperswithcode.com/paper/cnn-features-are-also-great-at-unsupervised |
Repo | https://github.com/jorisguerin/toolClustering_dataset |
Framework | none |
Referenceless Quality Estimation for Natural Language Generation
Title | Referenceless Quality Estimation for Natural Language Generation |
Authors | Ondřej Dušek, Jekaterina Novikova, Verena Rieser |
Abstract | Traditional automatic evaluation measures for natural language generation (NLG) use costly human-authored references to estimate the quality of a system output. In this paper, we propose a referenceless quality estimation (QE) approach based on recurrent neural networks, which predicts a quality score for a NLG system output by comparing it to the source meaning representation only. Our method outperforms traditional metrics and a constant baseline in most respects; we also show that synthetic data helps to increase correlation results by 21% compared to the base system. Our results are comparable to results obtained in similar QE tasks despite the more challenging setting. |
Tasks | Text Generation |
Published | 2017-08-05 |
URL | http://arxiv.org/abs/1708.01759v1 |
http://arxiv.org/pdf/1708.01759v1.pdf | |
PWC | https://paperswithcode.com/paper/referenceless-quality-estimation-for-natural |
Repo | https://github.com/tuetschek/ratpred |
Framework | none |
Learning Compact Geometric Features
Title | Learning Compact Geometric Features |
Authors | Marc Khoury, Qian-Yi Zhou, Vladlen Koltun |
Abstract | We present an approach to learning features that represent the local geometry around a point in an unstructured point cloud. Such features play a central role in geometric registration, which supports diverse applications in robotics and 3D vision. Current state-of-the-art local features for unstructured point clouds have been manually crafted and none combines the desirable properties of precision, compactness, and robustness. We show that features with these properties can be learned from data, by optimizing deep networks that map high-dimensional histograms into low-dimensional Euclidean spaces. The presented approach yields a family of features, parameterized by dimension, that are both more compact and more accurate than existing descriptors. |
Tasks | |
Published | 2017-09-15 |
URL | http://arxiv.org/abs/1709.05056v1 |
http://arxiv.org/pdf/1709.05056v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-compact-geometric-features |
Repo | https://github.com/marckhoury/CGF |
Framework | none |
Wasserstein Auto-Encoders
Title | Wasserstein Auto-Encoders |
Authors | Ilya Tolstikhin, Olivier Bousquet, Sylvain Gelly, Bernhard Schoelkopf |
Abstract | We propose the Wasserstein Auto-Encoder (WAE)—a new algorithm for building a generative model of the data distribution. WAE minimizes a penalized form of the Wasserstein distance between the model distribution and the target distribution, which leads to a different regularizer than the one used by the Variational Auto-Encoder (VAE). This regularizer encourages the encoded training distribution to match the prior. We compare our algorithm with several other techniques and show that it is a generalization of adversarial auto-encoders (AAE). Our experiments show that WAE shares many of the properties of VAEs (stable training, encoder-decoder architecture, nice latent manifold structure) while generating samples of better quality, as measured by the FID score. |
Tasks | |
Published | 2017-11-05 |
URL | https://arxiv.org/abs/1711.01558v4 |
https://arxiv.org/pdf/1711.01558v4.pdf | |
PWC | https://paperswithcode.com/paper/wasserstein-auto-encoders |
Repo | https://github.com/pravn/wasserstein_autoencoders |
Framework | pytorch |
Fine-tuning Tree-LSTM for phrase-level sentiment classification on a Polish dependency treebank. Submission to PolEval task 2
Title | Fine-tuning Tree-LSTM for phrase-level sentiment classification on a Polish dependency treebank. Submission to PolEval task 2 |
Authors | Tomasz Korbak, Paulina Żak |
Abstract | We describe a variant of Child-Sum Tree-LSTM deep neural network (Tai et al, 2015) fine-tuned for working with dependency trees and morphologically rich languages using the example of Polish. Fine-tuning included applying a custom regularization technique (zoneout, described by (Krueger et al., 2016), and further adapted for Tree-LSTMs) as well as using pre-trained word embeddings enhanced with sub-word information (Bojanowski et al., 2016). The system was implemented in PyTorch and evaluated on phrase-level sentiment labeling task as part of the PolEval competition. |
Tasks | Sentiment Analysis, Word Embeddings |
Published | 2017-11-03 |
URL | http://arxiv.org/abs/1711.01985v1 |
http://arxiv.org/pdf/1711.01985v1.pdf | |
PWC | https://paperswithcode.com/paper/fine-tuning-tree-lstm-for-phrase-level |
Repo | https://github.com/tomekkorbak/treehopper |
Framework | pytorch |
High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs
Title | High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs |
Authors | Ting-Chun Wang, Ming-Yu Liu, Jun-Yan Zhu, Andrew Tao, Jan Kautz, Bryan Catanzaro |
Abstract | We present a new method for synthesizing high-resolution photo-realistic images from semantic label maps using conditional generative adversarial networks (conditional GANs). Conditional GANs have enabled a variety of applications, but the results are often limited to low-resolution and still far from realistic. In this work, we generate 2048x1024 visually appealing results with a novel adversarial loss, as well as new multi-scale generator and discriminator architectures. Furthermore, we extend our framework to interactive visual manipulation with two additional features. First, we incorporate object instance segmentation information, which enables object manipulations such as removing/adding objects and changing the object category. Second, we propose a method to generate diverse results given the same input, allowing users to edit the object appearance interactively. Human opinion studies demonstrate that our method significantly outperforms existing methods, advancing both the quality and the resolution of deep image synthesis and editing. |
Tasks | Conditional Image Generation, Image Generation, Image-to-Image Translation, Instance Segmentation, Semantic Segmentation |
Published | 2017-11-30 |
URL | http://arxiv.org/abs/1711.11585v2 |
http://arxiv.org/pdf/1711.11585v2.pdf | |
PWC | https://paperswithcode.com/paper/high-resolution-image-synthesis-and-semantic |
Repo | https://github.com/eyeswideopen/pix2pixHD-MED |
Framework | pytorch |
Statistical Parametric Speech Synthesis Incorporating Generative Adversarial Networks
Title | Statistical Parametric Speech Synthesis Incorporating Generative Adversarial Networks |
Authors | Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari |
Abstract | A method for statistical parametric speech synthesis incorporating generative adversarial networks (GANs) is proposed. Although powerful deep neural networks (DNNs) techniques can be applied to artificially synthesize speech waveform, the synthetic speech quality is low compared with that of natural speech. One of the issues causing the quality degradation is an over-smoothing effect often observed in the generated speech parameters. A GAN introduced in this paper consists of two neural networks: a discriminator to distinguish natural and generated samples, and a generator to deceive the discriminator. In the proposed framework incorporating the GANs, the discriminator is trained to distinguish natural and generated speech parameters, while the acoustic models are trained to minimize the weighted sum of the conventional minimum generation loss and an adversarial loss for deceiving the discriminator. Since the objective of the GANs is to minimize the divergence (i.e., distribution difference) between the natural and generated speech parameters, the proposed method effectively alleviates the over-smoothing effect on the generated speech parameters. We evaluated the effectiveness for text-to-speech and voice conversion, and found that the proposed method can generate more natural spectral parameters and $F_0$ than conventional minimum generation error training algorithm regardless its hyper-parameter settings. Furthermore, we investigated the effect of the divergence of various GANs, and found that a Wasserstein GAN minimizing the Earth-Mover’s distance works the best in terms of improving synthetic speech quality. |
Tasks | Speech Synthesis, Voice Conversion |
Published | 2017-09-23 |
URL | http://arxiv.org/abs/1709.08041v1 |
http://arxiv.org/pdf/1709.08041v1.pdf | |
PWC | https://paperswithcode.com/paper/statistical-parametric-speech-synthesis-1 |
Repo | https://github.com/rickyHong/GAN-TTS-repl2 |
Framework | pytorch |
DeepVO: Towards End-to-End Visual Odometry with Deep Recurrent Convolutional Neural Networks
Title | DeepVO: Towards End-to-End Visual Odometry with Deep Recurrent Convolutional Neural Networks |
Authors | Sen Wang, Ronald Clark, Hongkai Wen, Niki Trigoni |
Abstract | This paper studies monocular visual odometry (VO) problem. Most of existing VO algorithms are developed under a standard pipeline including feature extraction, feature matching, motion estimation, local optimisation, etc. Although some of them have demonstrated superior performance, they usually need to be carefully designed and specifically fine-tuned to work well in different environments. Some prior knowledge is also required to recover an absolute scale for monocular VO. This paper presents a novel end-to-end framework for monocular VO by using deep Recurrent Convolutional Neural Networks (RCNNs). Since it is trained and deployed in an end-to-end manner, it infers poses directly from a sequence of raw RGB images (videos) without adopting any module in the conventional VO pipeline. Based on the RCNNs, it not only automatically learns effective feature representation for the VO problem through Convolutional Neural Networks, but also implicitly models sequential dynamics and relations using deep Recurrent Neural Networks. Extensive experiments on the KITTI VO dataset show competitive performance to state-of-the-art methods, verifying that the end-to-end Deep Learning technique can be a viable complement to the traditional VO systems. |
Tasks | Monocular Visual Odometry, Motion Estimation, Visual Odometry |
Published | 2017-09-25 |
URL | http://arxiv.org/abs/1709.08429v1 |
http://arxiv.org/pdf/1709.08429v1.pdf | |
PWC | https://paperswithcode.com/paper/deepvo-towards-end-to-end-visual-odometry |
Repo | https://github.com/fshamshirdar/DeepVO |
Framework | pytorch |
Construction of a Japanese Word Similarity Dataset
Title | Construction of a Japanese Word Similarity Dataset |
Authors | Yuya Sakaizawa, Mamoru Komachi |
Abstract | An evaluation of distributed word representation is generally conducted using a word similarity task and/or a word analogy task. There are many datasets readily available for these tasks in English. However, evaluating distributed representation in languages that do not have such resources (e.g., Japanese) is difficult. Therefore, as a first step toward evaluating distributed representations in Japanese, we constructed a Japanese word similarity dataset. To the best of our knowledge, our dataset is the first resource that can be used to evaluate distributed representations in Japanese. Moreover, our dataset contains various parts of speech and includes rare words in addition to common words. |
Tasks | |
Published | 2017-03-17 |
URL | http://arxiv.org/abs/1703.05916v2 |
http://arxiv.org/pdf/1703.05916v2.pdf | |
PWC | https://paperswithcode.com/paper/construction-of-a-japanese-word-similarity |
Repo | https://github.com/kdrl/SCNE |
Framework | none |