Paper Group AWR 197
Towards Analyzing Semantic Robustness of Deep Neural Networks. Neural Network Distiller: A Python Package For DNN Compression Research. Fast and Robust Dynamic Hand Gesture Recognition via Key Frames Extraction and Feature Fusion. Adversarial Out-domain Examples for Generative Models. Multi-Agent Deep Reinforcement Learning for Large-scale Traffic …
Towards Analyzing Semantic Robustness of Deep Neural Networks
Title | Towards Analyzing Semantic Robustness of Deep Neural Networks |
Authors | Abdullah Hamdi, Bernard Ghanem |
Abstract | Despite the impressive performance of Deep Neural Networks (DNNs) on various vision tasks, they still exhibit erroneous high sensitivity toward semantic primitives (e.g. object pose). We propose a theoretically grounded analysis for DNNs robustness in the semantic space. We qualitatively analyze different DNNs semantic robustness by visualizing the DNN global behavior as semantic maps and observe interesting behavior of some DNNs. Since generating these semantic maps does not scale well with the dimensionality of the semantic space, we develop a bottom-up approach to detect robust regions of DNNs. To achieve this, We formalize the problem of finding robust semantic regions of the network as optimization of integral bounds and develop expressions for update directions of the region bounds. We use our developed formulations to quantitatively evaluate the semantic robustness of different famous network architectures. We show through extensive experimentation that several networks, though trained on the same dataset and while enjoying comparable accuracy, they do not necessarily perform similarly in semantic robustness. For example, InceptionV3 is more accurate despite being less semantically robust than ResNet50. We hope that this tool will serve as the first milestone towards understanding the semantic robustness of DNNs. |
Tasks | Adversarial Attack, Autonomous Driving, Image Classification |
Published | 2019-04-09 |
URL | https://arxiv.org/abs/1904.04621v2 |
https://arxiv.org/pdf/1904.04621v2.pdf | |
PWC | https://paperswithcode.com/paper/towards-analyzing-semantic-robustness-of-deep |
Repo | https://github.com/ajhamdi/semantic-robustness |
Framework | pytorch |
Neural Network Distiller: A Python Package For DNN Compression Research
Title | Neural Network Distiller: A Python Package For DNN Compression Research |
Authors | Neta Zmora, Guy Jacob, Lev Zlotnik, Bar Elharar, Gal Novik |
Abstract | This paper presents the philosophy, design and feature-set of Neural Network Distiller, an open-source Python package for DNN compression research. Distiller is a library of DNN compression algorithms implementations, with tools, tutorials and sample applications for various learning tasks. Its target users are both engineers and researchers, and the rich content is complemented by a design-for-extensibility to facilitate new research. Distiller is open-source and is available on Github at https://github.com/NervanaSystems/distiller. |
Tasks | |
Published | 2019-10-27 |
URL | https://arxiv.org/abs/1910.12232v1 |
https://arxiv.org/pdf/1910.12232v1.pdf | |
PWC | https://paperswithcode.com/paper/neural-network-distiller-a-python-package-for |
Repo | https://github.com/NervanaSystems/distiller |
Framework | pytorch |
Fast and Robust Dynamic Hand Gesture Recognition via Key Frames Extraction and Feature Fusion
Title | Fast and Robust Dynamic Hand Gesture Recognition via Key Frames Extraction and Feature Fusion |
Authors | Hao Tang, Hong Liu, Wei Xiao, Nicu Sebe |
Abstract | Gesture recognition is a hot topic in computer vision and pattern recognition, which plays a vitally important role in natural human-computer interface. Although great progress has been made recently, fast and robust hand gesture recognition remains an open problem, since the existing methods have not well balanced the performance and the efficiency simultaneously. To bridge it, this work combines image entropy and density clustering to exploit the key frames from hand gesture video for further feature extraction, which can improve the efficiency of recognition. Moreover, a feature fusion strategy is also proposed to further improve feature representation, which elevates the performance of recognition. To validate our approach in a “wild” environment, we also introduce two new datasets called HandGesture and Action3D datasets. Experiments consistently demonstrate that our strategy achieves competitive results on Northwestern University, Cambridge, HandGesture and Action3D hand gesture datasets. Our code and datasets will release at https://github.com/Ha0Tang/HandGestureRecognition. |
Tasks | Gesture Recognition, Hand Gesture Recognition, Hand-Gesture Recognition |
Published | 2019-01-15 |
URL | http://arxiv.org/abs/1901.04622v1 |
http://arxiv.org/pdf/1901.04622v1.pdf | |
PWC | https://paperswithcode.com/paper/fast-and-robust-dynamic-hand-gesture |
Repo | https://github.com/Ha0Tang/HandGestureRecognition |
Framework | none |
Adversarial Out-domain Examples for Generative Models
Title | Adversarial Out-domain Examples for Generative Models |
Authors | Dario Pasquini, Marco Mingione, Massimo Bernaschi |
Abstract | Deep generative models are rapidly becoming a common tool for researchers and developers. However, as exhaustively shown for the family of discriminative models, the test-time inference of deep neural networks cannot be fully controlled and erroneous behaviors can be induced by an attacker. In the present work, we show how a malicious user can force a pre-trained generator to reproduce arbitrary data instances by feeding it suitable adversarial inputs. Moreover, we show that these adversarial latent vectors can be shaped so as to be statistically indistinguishable from the set of genuine inputs. The proposed attack technique is evaluated with respect to various GAN images generators using different architectures, training processes and for both conditional and not-conditional setups. |
Tasks | Adversarial Attack, Image Generation |
Published | 2019-03-07 |
URL | https://arxiv.org/abs/1903.02926v2 |
https://arxiv.org/pdf/1903.02926v2.pdf | |
PWC | https://paperswithcode.com/paper/out-domain-examples-for-generative-models |
Repo | https://github.com/pasquini-dario/OutDomainExamples |
Framework | tf |
Multi-Agent Deep Reinforcement Learning for Large-scale Traffic Signal Control
Title | Multi-Agent Deep Reinforcement Learning for Large-scale Traffic Signal Control |
Authors | Tianshu Chu, Jie Wang, Lara Codecà, Zhaojian Li |
Abstract | Reinforcement learning (RL) is a promising data-driven approach for adaptive traffic signal control (ATSC) in complex urban traffic networks, and deep neural networks further enhance its learning power. However, centralized RL is infeasible for large-scale ATSC due to the extremely high dimension of the joint action space. Multi-agent RL (MARL) overcomes the scalability issue by distributing the global control to each local RL agent, but it introduces new challenges: now the environment becomes partially observable from the viewpoint of each local agent due to limited communication among agents. Most existing studies in MARL focus on designing efficient communication and coordination among traditional Q-learning agents. This paper presents, for the first time, a fully scalable and decentralized MARL algorithm for the state-of-the-art deep RL agent: advantage actor critic (A2C), within the context of ATSC. In particular, two methods are proposed to stabilize the learning procedure, by improving the observability and reducing the learning difficulty of each local agent. The proposed multi-agent A2C is compared against independent A2C and independent Q-learning algorithms, in both a large synthetic traffic grid and a large real-world traffic network of Monaco city, under simulated peak-hour traffic dynamics. Results demonstrate its optimality, robustness, and sample efficiency over other state-of-the-art decentralized MARL algorithms. |
Tasks | Q-Learning |
Published | 2019-03-11 |
URL | http://arxiv.org/abs/1903.04527v1 |
http://arxiv.org/pdf/1903.04527v1.pdf | |
PWC | https://paperswithcode.com/paper/multi-agent-deep-reinforcement-learning-for-2 |
Repo | https://github.com/cts198859/deeprl_signal_control |
Framework | tf |
Privacy-preserving Q-Learning with Functional Noise in Continuous State Spaces
Title | Privacy-preserving Q-Learning with Functional Noise in Continuous State Spaces |
Authors | Baoxiang Wang, Nidhi Hegde |
Abstract | We consider differentially private algorithms for reinforcement learning in continuous spaces, such that neighboring reward functions are indistinguishable. This protects the reward information from being exploited by methods such as inverse reinforcement learning. Existing studies that guarantee differential privacy are not extendable to infinite state spaces, as the noise level to ensure privacy will scale accordingly to infinity. Our aim is to protect the value function approximator, without regard to the number of states queried to the function. It is achieved by adding functional noise to the value function iteratively in the training. We show rigorous privacy guarantees by a series of analyses on the kernel of the noise space, the probabilistic bound of such noise samples, and the composition over the iterations. We gain insight into the utility analysis by proving the algorithm’s approximate optimality when the state space is discrete. Experiments corroborate our theoretical findings and show improvement over existing approaches. |
Tasks | Q-Learning |
Published | 2019-01-30 |
URL | https://arxiv.org/abs/1901.10634v3 |
https://arxiv.org/pdf/1901.10634v3.pdf | |
PWC | https://paperswithcode.com/paper/private-q-learning-with-functional-noise-in |
Repo | https://github.com/wangbx66/differentially-private-q-learning |
Framework | pytorch |
Multi-Hop Paragraph Retrieval for Open-Domain Question Answering
Title | Multi-Hop Paragraph Retrieval for Open-Domain Question Answering |
Authors | Yair Feldman, Ran El-Yaniv |
Abstract | This paper is concerned with the task of multi-hop open-domain Question Answering (QA). This task is particularly challenging since it requires the simultaneous performance of textual reasoning and efficient searching. We present a method for retrieving multiple supporting paragraphs, nested amidst a large knowledge base, which contain the necessary evidence to answer a given question. Our method iteratively retrieves supporting paragraphs by forming a joint vector representation of both a question and a paragraph. The retrieval is performed by considering contextualized sentence-level representations of the paragraphs in the knowledge source. Our method achieves state-of-the-art performance over two well-known datasets, SQuAD-Open and HotpotQA, which serve as our single- and multi-hop open-domain QA benchmarks, respectively. |
Tasks | Open-Domain Question Answering, Question Answering |
Published | 2019-06-15 |
URL | https://arxiv.org/abs/1906.06606v1 |
https://arxiv.org/pdf/1906.06606v1.pdf | |
PWC | https://paperswithcode.com/paper/multi-hop-paragraph-retrieval-for-open-domain |
Repo | https://github.com/yairf11/MUPPET |
Framework | tf |
The Regretful Agent: Heuristic-Aided Navigation through Progress Estimation
Title | The Regretful Agent: Heuristic-Aided Navigation through Progress Estimation |
Authors | Chih-Yao Ma, Zuxuan Wu, Ghassan AlRegib, Caiming Xiong, Zsolt Kira |
Abstract | As deep learning continues to make progress for challenging perception tasks, there is increased interest in combining vision, language, and decision-making. Specifically, the Vision and Language Navigation (VLN) task involves navigating to a goal purely from language instructions and visual information without explicit knowledge of the goal. Recent successful approaches have made in-roads in achieving good success rates for this task but rely on beam search, which thoroughly explores a large number of trajectories and is unrealistic for applications such as robotics. In this paper, inspired by the intuition of viewing the problem as search on a navigation graph, we propose to use a progress monitor developed in prior work as a learnable heuristic for search. We then propose two modules incorporated into an end-to-end architecture: 1) A learned mechanism to perform backtracking, which decides whether to continue moving forward or roll back to a previous state (Regret Module) and 2) A mechanism to help the agent decide which direction to go next by showing directions that are visited and their associated progress estimate (Progress Marker). Combined, the proposed approach significantly outperforms current state-of-the-art methods using greedy action selection, with 5% absolute improvement on the test server in success rates, and more importantly 8% on success rates normalized by the path length. Our code is available at https://github.com/chihyaoma/regretful-agent . |
Tasks | Decision Making, Vision-Language Navigation, Visual Navigation |
Published | 2019-03-05 |
URL | http://arxiv.org/abs/1903.01602v1 |
http://arxiv.org/pdf/1903.01602v1.pdf | |
PWC | https://paperswithcode.com/paper/the-regretful-agent-heuristic-aided |
Repo | https://github.com/chihyaoma/regretful-agent |
Framework | pytorch |
Marathon Environments: Multi-Agent Continuous Control Benchmarks in a Modern Video Game Engine
Title | Marathon Environments: Multi-Agent Continuous Control Benchmarks in a Modern Video Game Engine |
Authors | Joe Booth, Jackson Booth |
Abstract | Recent advances in deep reinforcement learning in the paradigm of locomotion using continuous control have raised the interest of game makers for the potential of digital actors using active ragdoll. Currently, the available options to develop these ideas are either researchers’ limited codebase or proprietary closed systems. We present Marathon Environments, a suite of open source, continuous control benchmarks implemented on the Unity game engine, using the Unity ML- Agents Toolkit. We demonstrate through these benchmarks that continuous control research is transferable to a commercial game engine. Furthermore, we exhibit the robustness of these environments by reproducing advanced continuous control research, such as learning to walk, run and backflip from motion capture data; learning to navigate complex terrains; and by implementing a video game input control system. We show further robustness by training with alternative algorithms found in OpenAI.Baselines. Finally, we share strategies for significantly reducing the training time. |
Tasks | Continuous Control, Motion Capture |
Published | 2019-02-25 |
URL | http://arxiv.org/abs/1902.09097v1 |
http://arxiv.org/pdf/1902.09097v1.pdf | |
PWC | https://paperswithcode.com/paper/marathon-environments-multi-agent-continuous |
Repo | https://github.com/Unity-Technologies/marathon-envs |
Framework | none |
DIANet: Dense-and-Implicit Attention Network
Title | DIANet: Dense-and-Implicit Attention Network |
Authors | Zhongzhan Huang, Senwei Liang, Mingfu Liang, Haizhao Yang |
Abstract | Attention networks have successfully boosted the performance in various vision problems. Previous works lay emphasis on designing a new attention module and individually plug them into the networks. Our paper proposes a novel-and-simple framework that shares an attention module throughout different network layers to encourage the integration of layer-wise information and this parameter-sharing module is referred as Dense-and-Implicit-Attention (DIA) unit. Many choices of modules can be used in the DIA unit. Since Long Short Term Memory (LSTM) has a capacity of capturing long-distance dependency, we focus on the case when the DIA unit is the modified LSTM (refer as DIA-LSTM). Experiments on benchmark datasets show that the DIA-LSTM unit is capable of emphasizing layer-wise feature interrelation and leads to significant improvement of image classification accuracy. We further empirically show that the DIA-LSTM has a strong regularization ability on stabilizing the training of deep networks by the experiments with the removal of skip connections or Batch Normalization in the whole residual network. The code is released at https://github.com/gbup-group/DIANet. |
Tasks | Image Classification |
Published | 2019-05-25 |
URL | https://arxiv.org/abs/1905.10671v2 |
https://arxiv.org/pdf/1905.10671v2.pdf | |
PWC | https://paperswithcode.com/paper/dianet-dense-and-implicit-attention-network |
Repo | https://github.com/osmr/imgclsmob |
Framework | mxnet |
Neuropathic Pain Diagnosis Simulator for Causal Discovery Algorithm Evaluation
Title | Neuropathic Pain Diagnosis Simulator for Causal Discovery Algorithm Evaluation |
Authors | Ruibo Tu, Kun Zhang, Bo Christer Bertilson, Hedvig Kjellström, Cheng Zhang |
Abstract | Discovery of causal relations from observational data is essential for many disciplines of science and real-world applications. However, unlike other machine learning algorithms, whose development has been greatly fostered by a large amount of available benchmark datasets, causal discovery algorithms are notoriously difficult to be systematically evaluated because few datasets with known ground-truth causal relations are available. In this work, we handle the problem of evaluating causal discovery algorithms by building a flexible simulator in the medical setting. We develop a neuropathic pain diagnosis simulator, inspired by the fact that the biological processes of neuropathic pathophysiology are well studied with well-understood causal influences. Our simulator exploits the causal graph of the neuropathic pain pathology and its parameters in the generator are estimated from real-life patient cases. We show that the data generated from our simulator have similar statistics as real-world data. As a clear advantage, the simulator can produce infinite samples without jeopardizing the privacy of real-world patients. Our simulator provides a natural tool for evaluating various types of causal discovery algorithms, including those to deal with practical issues in causal discovery, such as unknown confounders, selection bias, and missing data. Using our simulator, we have evaluated extensively causal discovery algorithms under various settings. |
Tasks | Causal Discovery |
Published | 2019-06-04 |
URL | https://arxiv.org/abs/1906.01732v3 |
https://arxiv.org/pdf/1906.01732v3.pdf | |
PWC | https://paperswithcode.com/paper/neuropathic-pain-diagnosis-simulator-for |
Repo | https://github.com/TURuibo/Neuropathic-Pain-Diagnosis-Simulator |
Framework | none |
Emergence of Exploratory Look-Around Behaviors through Active Observation Completion
Title | Emergence of Exploratory Look-Around Behaviors through Active Observation Completion |
Authors | Santhosh K. Ramakrishnan, Dinesh Jayaraman, Kristen Grauman |
Abstract | Standard computer vision systems assume access to intelligently captured inputs (e.g., photos from a human photographer), yet autonomously capturing good observations is a major challenge in itself. We address the problem of learning to look around: how can an agent learn to acquire informative visual observations? We propose a reinforcement learning solution, where the agent is rewarded for reducing its uncertainty about the unobserved portions of its environment. Specifically, the agent is trained to select a short sequence of glimpses after which it must infer the appearance of its full environment. To address the challenge of sparse rewards, we further introduce sidekick policy learning, which exploits the asymmetry in observability between training and test time. The proposed methods learn observation policies that not only perform the completion task for which they are trained, but also generalize to exhibit useful “look-around” behavior for a range of active perception tasks. |
Tasks | Active Observation Completion, Observation Completion |
Published | 2019-06-27 |
URL | https://arxiv.org/abs/1906.11407v1 |
https://arxiv.org/pdf/1906.11407v1.pdf | |
PWC | https://paperswithcode.com/paper/emergence-of-exploratory-look-around-1 |
Repo | https://github.com/srama2512/sidekicks |
Framework | pytorch |
A Debiased MDI Feature Importance Measure for Random Forests
Title | A Debiased MDI Feature Importance Measure for Random Forests |
Authors | Xiao Li, Yu Wang, Sumanta Basu, Karl Kumbier, Bin Yu |
Abstract | Tree ensembles such as Random Forests have achieved impressive empirical success across a wide variety of applications. To understand how these models make predictions, people routinely turn to feature importance measures calculated from tree ensembles. It has long been known that Mean Decrease Impurity (MDI), one of the most widely used measures of feature importance, incorrectly assigns high importance to noisy features, leading to systematic bias in feature selection. In this paper, we address the feature selection bias of MDI from both theoretical and methodological perspectives. Based on the original definition of MDI by Breiman et al. for a single tree, we derive a tight non-asymptotic bound on the expected bias of MDI importance of noisy features, showing that deep trees have higher (expected) feature selection bias than shallow ones. However, it is not clear how to reduce the bias of MDI using its existing analytical expression. We derive a new analytical expression for MDI, and based on this new expression, we are able to propose a debiased MDI feature importance measure using out-of-bag samples, called MDI-oob. For both the simulated data and a genomic ChIP dataset, MDI-oob achieves state-of-the-art performance in feature selection from Random Forests for both deep and shallow trees. |
Tasks | Feature Importance, Feature Selection |
Published | 2019-06-26 |
URL | https://arxiv.org/abs/1906.10845v2 |
https://arxiv.org/pdf/1906.10845v2.pdf | |
PWC | https://paperswithcode.com/paper/a-debiased-mdi-feature-importance-measure-for |
Repo | https://github.com/nalzok/tree.interpreter |
Framework | none |
Asking Easy Questions: A User-Friendly Approach to Active Reward Learning
Title | Asking Easy Questions: A User-Friendly Approach to Active Reward Learning |
Authors | Erdem Bıyık, Malayandi Palan, Nicholas C. Landolfi, Dylan P. Losey, Dorsa Sadigh |
Abstract | Robots can learn the right reward function by querying a human expert. Existing approaches attempt to choose questions where the robot is most uncertain about the human’s response; however, they do not consider how easy it will be for the human to answer! In this paper we explore an information gain formulation for optimally selecting questions that naturally account for the human’s ability to answer. Our approach identifies questions that optimize the trade-off between robot and human uncertainty, and determines when these questions become redundant or costly. Simulations and a user study show our method not only produces easy questions, but also ultimately results in faster reward learning. |
Tasks | |
Published | 2019-10-10 |
URL | https://arxiv.org/abs/1910.04365v1 |
https://arxiv.org/pdf/1910.04365v1.pdf | |
PWC | https://paperswithcode.com/paper/asking-easy-questions-a-user-friendly |
Repo | https://github.com/Stanford-ILIAD/easy-active-learning |
Framework | none |
Multimodal Image Outpainting With Regularized Normalized Diversification
Title | Multimodal Image Outpainting With Regularized Normalized Diversification |
Authors | Lingzhi Zhang, Jiancong Wang, Jianbo Shi |
Abstract | In this paper, we study the problem of generating a set ofrealistic and diverse backgrounds when given only a smallforeground region. We refer to this task as image outpaint-ing. The technical challenge of this task is to synthesize notonly plausible but also diverse image outputs. Traditionalgenerative adversarial networks suffer from mode collapse.While recent approaches propose to maximize orpreserve the pairwise distance between generated sampleswith respect to their latent distance, they do not explicitlyprevent the diverse samples of different conditional inputsfrom collapsing. Therefore, we propose a new regulariza-tion method to encourage diverse sampling in conditionalsynthesis. In addition, we propose a feature pyramid dis-criminator to improve the image quality. Our experimen-tal results show that our model can produce more diverseimages without sacrificing visual quality compared to state-of-the-arts approaches in both the CelebA face dataset and the Cityscape scene dataset. |
Tasks | Image Outpainting |
Published | 2019-10-25 |
URL | https://arxiv.org/abs/1910.11481v1 |
https://arxiv.org/pdf/1910.11481v1.pdf | |
PWC | https://paperswithcode.com/paper/multimodal-image-outpainting-with-regularized |
Repo | https://github.com/owenzlz/Diverse_Outpaint |
Framework | pytorch |