Paper Group AWR 342
Autoregressive Quantile Networks for Generative Modeling. Guiding Deep Learning System Testing using Surprise Adequacy. Deep Randomized Ensembles for Metric Learning. MEAL: Multi-Model Ensemble via Adversarial Learning. Temporal Pattern Attention for Multivariate Time Series Forecasting. The Global Anchor Method for Quantifying Linguistic Shifts an …
Autoregressive Quantile Networks for Generative Modeling
Title | Autoregressive Quantile Networks for Generative Modeling |
Authors | Georg Ostrovski, Will Dabney, Rémi Munos |
Abstract | We introduce autoregressive implicit quantile networks (AIQN), a fundamentally different approach to generative modeling than those commonly used, that implicitly captures the distribution using quantile regression. AIQN is able to achieve superior perceptual quality and improvements in evaluation metrics, without incurring a loss of sample diversity. The method can be applied to many existing models and architectures. In this work we extend the PixelCNN model with AIQN and demonstrate results on CIFAR-10 and ImageNet using Inception score, FID, non-cherry-picked samples, and inpainting results. We consistently observe that AIQN yields a highly stable algorithm that improves perceptual quality while maintaining a highly diverse distribution. |
Tasks | |
Published | 2018-06-14 |
URL | http://arxiv.org/abs/1806.05575v1 |
http://arxiv.org/pdf/1806.05575v1.pdf | |
PWC | https://paperswithcode.com/paper/autoregressive-quantile-networks-for |
Repo | https://github.com/SSS135/aiqn-vae |
Framework | pytorch |
Guiding Deep Learning System Testing using Surprise Adequacy
Title | Guiding Deep Learning System Testing using Surprise Adequacy |
Authors | Jinhan Kim, Robert Feldt, Shin Yoo |
Abstract | Deep Learning (DL) systems are rapidly being adopted in safety and security critical domains, urgently calling for ways to test their correctness and robustness. Testing of DL systems has traditionally relied on manual collection and labelling of data. Recently, a number of coverage criteria based on neuron activation values have been proposed. These criteria essentially count the number of neurons whose activation during the execution of a DL system satisfied certain properties, such as being above predefined thresholds. However, existing coverage criteria are not sufficiently fine grained to capture subtle behaviours exhibited by DL systems. Moreover, evaluations have focused on showing correlation between adversarial examples and proposed criteria rather than evaluating and guiding their use for actual testing of DL systems. We propose a novel test adequacy criterion for testing of DL systems, called Surprise Adequacy for Deep Learning Systems (SADL), which is based on the behaviour of DL systems with respect to their training data. We measure the surprise of an input as the difference in DL system’s behaviour between the input and the training data (i.e., what was learnt during training), and subsequently develop this as an adequacy criterion: a good test input should be sufficiently but not overtly surprising compared to training data. Empirical evaluation using a range of DL systems from simple image classifiers to autonomous driving car platforms shows that systematic sampling of inputs based on their surprise can improve classification accuracy of DL systems against adversarial examples by up to 77.5% via retraining. |
Tasks | Autonomous Driving |
Published | 2018-08-25 |
URL | http://arxiv.org/abs/1808.08444v1 |
http://arxiv.org/pdf/1808.08444v1.pdf | |
PWC | https://paperswithcode.com/paper/guiding-deep-learning-system-testing-using |
Repo | https://github.com/lcckkkhaha/Suprise-Adequacy-Implementation |
Framework | none |
Deep Randomized Ensembles for Metric Learning
Title | Deep Randomized Ensembles for Metric Learning |
Authors | Hong Xuan, Richard Souvenir, Robert Pless |
Abstract | Learning embedding functions, which map semantically related inputs to nearby locations in a feature space supports a variety of classification and information retrieval tasks. In this work, we propose a novel, generalizable and fast method to define a family of embedding functions that can be used as an ensemble to give improved results. Each embedding function is learned by randomly bagging the training labels into small subsets. We show experimentally that these embedding ensembles create effective embedding functions. The ensemble output defines a metric space that improves state of the art performance for image retrieval on CUB-200-2011, Cars-196, In-Shop Clothes Retrieval and VehicleID. |
Tasks | Image Retrieval, Information Retrieval, Metric Learning |
Published | 2018-08-13 |
URL | http://arxiv.org/abs/1808.04469v2 |
http://arxiv.org/pdf/1808.04469v2.pdf | |
PWC | https://paperswithcode.com/paper/deep-randomized-ensembles-for-metric-learning |
Repo | https://github.com/littleredxh/DREML |
Framework | pytorch |
MEAL: Multi-Model Ensemble via Adversarial Learning
Title | MEAL: Multi-Model Ensemble via Adversarial Learning |
Authors | Zhiqiang Shen, Zhankui He, Xiangyang Xue |
Abstract | Often the best performing deep neural models are ensembles of multiple base-level networks. Unfortunately, the space required to store these many networks, and the time required to execute them at test-time, prohibits their use in applications where test sets are large (e.g., ImageNet). In this paper, we present a method for compressing large, complex trained ensembles into a single network, where knowledge from a variety of trained deep neural networks (DNNs) is distilled and transferred to a single DNN. In order to distill diverse knowledge from different trained (teacher) models, we propose to use adversarial-based learning strategy where we define a block-wise training loss to guide and optimize the predefined student network to recover the knowledge in teacher models, and to promote the discriminator network to distinguish teacher vs. student features simultaneously. The proposed ensemble method (MEAL) of transferring distilled knowledge with adversarial learning exhibits three important advantages: (1) the student network that learns the distilled knowledge with discriminators is optimized better than the original model; (2) fast inference is realized by a single forward pass, while the performance is even better than traditional ensembles from multi-original models; (3) the student network can learn the distilled knowledge from a teacher model that has arbitrary structures. Extensive experiments on CIFAR-10/100, SVHN and ImageNet datasets demonstrate the effectiveness of our MEAL method. On ImageNet, our ResNet-50 based MEAL achieves top-1/5 21.79%/5.99% val error, which outperforms the original model by 2.06%/1.14%. Code and models are available at: https://github.com/AaronHeee/MEAL |
Tasks | |
Published | 2018-12-06 |
URL | https://arxiv.org/abs/1812.02425v2 |
https://arxiv.org/pdf/1812.02425v2.pdf | |
PWC | https://paperswithcode.com/paper/meal-multi-model-ensemble-via-adversarial |
Repo | https://github.com/AaronHeee/MEAL |
Framework | pytorch |
Temporal Pattern Attention for Multivariate Time Series Forecasting
Title | Temporal Pattern Attention for Multivariate Time Series Forecasting |
Authors | Shun-Yao Shih, Fan-Keng Sun, Hung-yi Lee |
Abstract | Forecasting multivariate time series data, such as prediction of electricity consumption, solar power production, and polyphonic piano pieces, has numerous valuable applications. However, complex and non-linear interdependencies between time steps and series complicate the task. To obtain accurate prediction, it is crucial to model long-term dependency in time series data, which can be achieved to some good extent by recurrent neural network (RNN) with attention mechanism. Typical attention mechanism reviews the information at each previous time step and selects the relevant information to help generate the outputs, but it fails to capture the temporal patterns across multiple time steps. In this paper, we propose to use a set of filters to extract time-invariant temporal patterns, which is similar to transforming time series data into its “frequency domain”. Then we proposed a novel attention mechanism to select relevant time series, and use its “frequency domain” information for forecasting. We applied the proposed model on several real-world tasks and achieved state-of-the-art performance in all of them with only one exception. |
Tasks | Multivariate Time Series Forecasting, Time Series, Time Series Forecasting |
Published | 2018-09-12 |
URL | https://arxiv.org/abs/1809.04206v3 |
https://arxiv.org/pdf/1809.04206v3.pdf | |
PWC | https://paperswithcode.com/paper/temporal-pattern-attention-for-multivariate |
Repo | https://github.com/abinashsinha330/Air-Pollution-Forecasting-using-Recurrent-Neural-Networks |
Framework | tf |
The Global Anchor Method for Quantifying Linguistic Shifts and Domain Adaptation
Title | The Global Anchor Method for Quantifying Linguistic Shifts and Domain Adaptation |
Authors | Zi Yin, Vin Sachidananda, Balaji Prabhakar |
Abstract | Language is dynamic, constantly evolving and adapting with respect to time, domain or topic. The adaptability of language is an active research area, where researchers discover social, cultural and domain-specific changes in language using distributional tools such as word embeddings. In this paper, we introduce the global anchor method for detecting corpus-level language shifts. We show both theoretically and empirically that the global anchor method is equivalent to the alignment method, a widely-used method for comparing word embeddings, in terms of detecting corpus-level language shifts. Despite their equivalence in terms of detection abilities, we demonstrate that the global anchor method is superior in terms of applicability as it can compare embeddings of different dimensionalities. Furthermore, the global anchor method has implementation and parallelization advantages. We show that the global anchor method reveals fine structures in the evolution of language and domain adaptation. When combined with the graph Laplacian technique, the global anchor method recovers the evolution trajectory and domain clustering of disparate text corpora. |
Tasks | Domain Adaptation, Word Embeddings |
Published | 2018-12-12 |
URL | http://arxiv.org/abs/1812.10382v1 |
http://arxiv.org/pdf/1812.10382v1.pdf | |
PWC | https://paperswithcode.com/paper/the-global-anchor-method-for-quantifying |
Repo | https://github.com/ziyin-dl/global-anchor-method |
Framework | none |
Adapting multi-armed bandits policies to contextual bandits scenarios
Title | Adapting multi-armed bandits policies to contextual bandits scenarios |
Authors | David Cortes |
Abstract | This work explores adaptations of successful multi-armed bandits policies to the online contextual bandits scenario with binary rewards using binary classification algorithms such as logistic regression as black-box oracles. Some of these adaptations are achieved through bootstrapping or approximate bootstrapping, while others rely on other forms of randomness, resulting in more scalable approaches than previous works, and the ability to work with any type of classification algorithm. In particular, the Adaptive-Greedy algorithm shows a lot of promise, in many cases achieving better performance than upper confidence bound and Thompson sampling strategies, at the expense of more hyperparameters to tune. |
Tasks | Multi-Armed Bandits |
Published | 2018-11-11 |
URL | https://arxiv.org/abs/1811.04383v2 |
https://arxiv.org/pdf/1811.04383v2.pdf | |
PWC | https://paperswithcode.com/paper/adapting-multi-armed-bandits-policies-to |
Repo | https://github.com/david-cortes/contextualbandits |
Framework | none |
Capsules for Object Segmentation
Title | Capsules for Object Segmentation |
Authors | Rodney LaLonde, Ulas Bagci |
Abstract | Convolutional neural networks (CNNs) have shown remarkable results over the last several years for a wide range of computer vision tasks. A new architecture recently introduced by Sabour et al., referred to as a capsule networks with dynamic routing, has shown great initial results for digit recognition and small image classification. The success of capsule networks lies in their ability to preserve more information about the input by replacing max-pooling layers with convolutional strides and dynamic routing, allowing for preservation of part-whole relationships in the data. This preservation of the input is demonstrated by reconstructing the input from the output capsule vectors. Our work expands the use of capsule networks to the task of object segmentation for the first time in the literature. We extend the idea of convolutional capsules with locally-connected routing and propose the concept of deconvolutional capsules. Further, we extend the masked reconstruction to reconstruct the positive input class. The proposed convolutional-deconvolutional capsule network, called SegCaps, shows strong results for the task of object segmentation with substantial decrease in parameter space. As an example application, we applied the proposed SegCaps to segment pathological lungs from low dose CT scans and compared its accuracy and efficiency with other U-Net-based architectures. SegCaps is able to handle large image sizes (512 x 512) as opposed to baseline capsules (typically less than 32 x 32). The proposed SegCaps reduced the number of parameters of U-Net architecture by 95.4% while still providing a better segmentation accuracy. |
Tasks | Semantic Segmentation |
Published | 2018-04-11 |
URL | http://arxiv.org/abs/1804.04241v1 |
http://arxiv.org/pdf/1804.04241v1.pdf | |
PWC | https://paperswithcode.com/paper/capsules-for-object-segmentation |
Repo | https://github.com/Cheng-Lin-Li/SegCaps |
Framework | tf |
Fully Decentralized Multi-Agent Reinforcement Learning with Networked Agents
Title | Fully Decentralized Multi-Agent Reinforcement Learning with Networked Agents |
Authors | Kaiqing Zhang, Zhuoran Yang, Han Liu, Tong Zhang, Tamer Başar |
Abstract | We consider the problem of \emph{fully decentralized} multi-agent reinforcement learning (MARL), where the agents are located at the nodes of a time-varying communication network. Specifically, we assume that the reward functions of the agents might correspond to different tasks, and are only known to the corresponding agent. Moreover, each agent makes individual decisions based on both the information observed locally and the messages received from its neighbors over the network. Within this setting, the collective goal of the agents is to maximize the globally averaged return over the network through exchanging information with their neighbors. To this end, we propose two decentralized actor-critic algorithms with function approximation, which are applicable to large-scale MARL problems where both the number of states and the number of agents are massively large. Under the decentralized structure, the actor step is performed individually by each agent with no need to infer the policies of others. For the critic step, we propose a consensus update via communication over the network. Our algorithms are fully incremental and can be implemented in an online fashion. Convergence analyses of the algorithms are provided when the value functions are approximated within the class of linear functions. Extensive simulation results with both linear and nonlinear function approximations are presented to validate the proposed algorithms. Our work appears to be the first study of fully decentralized MARL algorithms for networked agents with function approximation, with provable convergence guarantees. |
Tasks | Multi-agent Reinforcement Learning |
Published | 2018-02-23 |
URL | http://arxiv.org/abs/1802.08757v2 |
http://arxiv.org/pdf/1802.08757v2.pdf | |
PWC | https://paperswithcode.com/paper/fully-decentralized-multi-agent-reinforcement |
Repo | https://github.com/MUmarJaved/MultiAgent-Distributed-Reinforcement-Learning |
Framework | tf |
Deep Reinforcement Learning
Title | Deep Reinforcement Learning |
Authors | Yuxi Li |
Abstract | We discuss deep reinforcement learning in an overview style. We draw a big picture, filled with details. We discuss six core elements, six important mechanisms, and twelve applications, focusing on contemporary work, and in historical contexts. We start with background of artificial intelligence, machine learning, deep learning, and reinforcement learning (RL), with resources. Next we discuss RL core elements, including value function, policy, reward, model, exploration vs. exploitation, and representation. Then we discuss important mechanisms for RL, including attention and memory, unsupervised learning, hierarchical RL, multi-agent RL, relational RL, and learning to learn. After that, we discuss RL applications, including games, robotics, natural language processing (NLP), computer vision, finance, business management, healthcare, education, energy, transportation, computer systems, and, science, engineering, and art. Finally we summarize briefly, discuss challenges and opportunities, and close with an epilogue. |
Tasks | |
Published | 2018-10-15 |
URL | http://arxiv.org/abs/1810.06339v1 |
http://arxiv.org/pdf/1810.06339v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-reinforcement-learning |
Repo | https://github.com/google/dopamine |
Framework | tf |
Machine Teaching of Active Sequential Learners
Title | Machine Teaching of Active Sequential Learners |
Authors | Tomi Peltola, Mustafa Mert Çelikok, Pedram Daee, Samuel Kaski |
Abstract | Machine teaching addresses the problem of finding the best training data that can guide a learning algorithm to a target model with minimal effort. In conventional settings, a teacher provides data that are consistent with the true data distribution. However, for sequential learners which actively choose their queries, such as multi-armed bandits and active learners, the teacher can only provide responses to the learner’s queries, not design the full data. In this setting, consistent teachers can be sub-optimal for finite horizons. We formulate this sequential teaching problem, which current techniques in machine teaching do not address, as a Markov decision process, with the dynamics nesting a model of the learner and the actions being the teacher’s responses. Furthermore, we address the complementary problem of learning from a teacher that plans: to recognise the teaching intent of the responses, the learner is endowed with a model of the teacher. We test the formulation with multi-armed bandit learners in simulated experiments and a user study. The results show that learning is improved by (i) planning teaching and (ii) the learner having a model of the teacher. The approach gives tools to taking into account strategic (planning) behaviour of users of interactive intelligent systems, such as recommendation engines, by considering them as boundedly optimal teachers. |
Tasks | Multi-Armed Bandits, Probabilistic Programming |
Published | 2018-09-08 |
URL | https://arxiv.org/abs/1809.02869v3 |
https://arxiv.org/pdf/1809.02869v3.pdf | |
PWC | https://paperswithcode.com/paper/modelling-users-theory-of-ais-mind-in |
Repo | https://github.com/AaltoPML/machine-teaching-of-active-sequential-learners |
Framework | pytorch |
On-line Adaptative Curriculum Learning for GANs
Title | On-line Adaptative Curriculum Learning for GANs |
Authors | Thang Doan, Joao Monteiro, Isabela Albuquerque, Bogdan Mazoure, Audrey Durand, Joelle Pineau, R Devon Hjelm |
Abstract | Generative Adversarial Networks (GANs) can successfully approximate a probability distribution and produce realistic samples. However, open questions such as sufficient convergence conditions and mode collapse still persist. In this paper, we build on existing work in the area by proposing a novel framework for training the generator against an ensemble of discriminator networks, which can be seen as a one-student/multiple-teachers setting. We formalize this problem within the full-information adversarial bandit framework, where we evaluate the capability of an algorithm to select mixtures of discriminators for providing the generator with feedback during learning. To this end, we propose a reward function which reflects the progress made by the generator and dynamically update the mixture weights allocated to each discriminator. We also draw connections between our algorithm and stochastic optimization methods and then show that existing approaches using multiple discriminators in literature can be recovered from our framework. We argue that less expressive discriminators are smoother and have a general coarse grained view of the modes map, which enforces the generator to cover a wide portion of the data distribution support. On the other hand, highly expressive discriminators ensure samples quality. Finally, experimental results show that our approach improves samples quality and diversity over existing baselines by effectively learning a curriculum. These results also support the claim that weaker discriminators have higher entropy improving modes coverage. Keywords: multiple discriminators, curriculum learning, multiple resolutions discriminators, multi-armed bandits, generative adversarial networks, smooth discriminators, multi-discriminator gan training, multiple experts. |
Tasks | Multi-Armed Bandits, Stochastic Optimization |
Published | 2018-07-31 |
URL | http://arxiv.org/abs/1808.00020v6 |
http://arxiv.org/pdf/1808.00020v6.pdf | |
PWC | https://paperswithcode.com/paper/on-line-adaptative-curriculum-learning-for |
Repo | https://github.com/Byte7/Adaptative-Curriculum-GAN-keras |
Framework | none |
PAC-Bayes Control: Learning Policies that Provably Generalize to Novel Environments
Title | PAC-Bayes Control: Learning Policies that Provably Generalize to Novel Environments |
Authors | Anirudha Majumdar, Alec Farid, Anoopkumar Sonar |
Abstract | Our goal is to learn control policies for robots that provably generalize well to novel environments given a dataset of example environments. The key technical idea behind our approach is to leverage tools from generalization theory in machine learning by exploiting a precise analogy (which we present in the form of a reduction) between generalization of control policies to novel environments and generalization of hypotheses in the supervised learning setting. In particular, we utilize the Probably Approximately Correct (PAC)-Bayes framework, which allows us to obtain upper bounds that hold with high probability on the expected cost of (stochastic) control policies across novel environments. We propose policy learning algorithms that explicitly seek to minimize this upper bound. The corresponding optimization problem can be solved using convex optimization (Relative Entropy Programming in particular) in the setting where we are optimizing over a finite policy space. In the more general setting of continuously parameterized policies (e.g., neural network policies), we minimize this upper bound using stochastic gradient descent. We present simulated results of our approach applied to learning (1) reactive obstacle avoidance policies and (2) neural network-based grasping policies. We also present hardware results for the Parrot Swing drone navigating through different obstacle environments. Our examples demonstrate the potential of our approach to provide strong generalization guarantees for robotic systems with continuous state and action spaces, complicated (e.g., nonlinear) dynamics, rich sensory inputs (e.g., depth images), and neural network-based policies. |
Tasks | |
Published | 2018-06-11 |
URL | https://arxiv.org/abs/1806.04225v4 |
https://arxiv.org/pdf/1806.04225v4.pdf | |
PWC | https://paperswithcode.com/paper/pac-bayes-control-synthesizing-controllers |
Repo | https://github.com/irom-lab/PAC-Bayes-Control |
Framework | none |
Graph R-CNN for Scene Graph Generation
Title | Graph R-CNN for Scene Graph Generation |
Authors | Jianwei Yang, Jiasen Lu, Stefan Lee, Dhruv Batra, Devi Parikh |
Abstract | We propose a novel scene graph generation model called Graph R-CNN, that is both effective and efficient at detecting objects and their relations in images. Our model contains a Relation Proposal Network (RePN) that efficiently deals with the quadratic number of potential relations between objects in an image. We also propose an attentional Graph Convolutional Network (aGCN) that effectively captures contextual information between objects and relations. Finally, we introduce a new evaluation metric that is more holistic and realistic than existing metrics. We report state-of-the-art performance on scene graph generation as evaluated using both existing and our proposed metrics. |
Tasks | Graph Generation, Scene Graph Generation |
Published | 2018-08-01 |
URL | http://arxiv.org/abs/1808.00191v1 |
http://arxiv.org/pdf/1808.00191v1.pdf | |
PWC | https://paperswithcode.com/paper/graph-r-cnn-for-scene-graph-generation |
Repo | https://github.com/jwyang/graph-rcnn.pytorch |
Framework | pytorch |
RGBiD-SLAM for Accurate Real-time Localisation and 3D Mapping
Title | RGBiD-SLAM for Accurate Real-time Localisation and 3D Mapping |
Authors | Daniel Gutierrez-Gomez, Jose J. Guerrero |
Abstract | In this paper we present a complete SLAM system for RGB-D cameras, namely RGB-iD SLAM. The presented approach is a dense direct SLAM method with the main characteristic of working with the depth maps in inverse depth parametrisation for the routines of dense alignment or keyframe fusion. The system consists in 2 CPU threads working in parallel, which share the use of the GPU for dense alignment and keyframe fusion routines. The first thread is a front-end operating at frame rate, which processes every incoming frame from the RGB-D sensor to compute the incremental odometry and integrate it in a keyframe which is changed periodically following a covisibility-based strategy. The second thread is a back-end which receives keyframes from the front-end. This thread is in charge of segmenting the keyframes based on their structure, describing them using Bags of Words, trying to find potential loop closures with previous keyframes, and in such case perform pose-graph optimisation for trajectory correction. In addition, our system allows is able to compute the odometry both with unregistered and registered depth maps, allowing to use customised calibrations of the RGB-D sensor. As a consequence in the paper we also propose a detailed calibration pipeline to compute customised calibrations for particular RGB-D cameras. The experiments with our approach in the TUM RGB-D benchmark datasets show results superior in accuracy to the state-of-the-art in many of the sequences. The code has been made available on-line for research purposes https://github.com/dangut/RGBiD-SLAM. |
Tasks | Calibration |
Published | 2018-07-22 |
URL | http://arxiv.org/abs/1807.08271v1 |
http://arxiv.org/pdf/1807.08271v1.pdf | |
PWC | https://paperswithcode.com/paper/rgbid-slam-for-accurate-real-time |
Repo | https://github.com/dangut/RGBiD-SLAM |
Framework | none |