Paper Group AWR 350
Differentially Private Bayesian Inference for Exponential Families. Evolving Multimodal Robot Behavior via Many Stepping Stones with the Combinatorial Multi-Objective Evolutionary Algorithm. Found Graph Data and Planted Vertex Covers. Deblending galaxy superpositions with branched generative adversarial networks. Scientific image rendering for spac …
Differentially Private Bayesian Inference for Exponential Families
Title | Differentially Private Bayesian Inference for Exponential Families |
Authors | Garrett Bernstein, Daniel Sheldon |
Abstract | The study of private inference has been sparked by growing concern regarding the analysis of data when it stems from sensitive sources. We present the first method for private Bayesian inference in exponential families that properly accounts for noise introduced by the privacy mechanism. It is efficient because it works only with sufficient statistics and not individual data. Unlike other methods, it gives properly calibrated posterior beliefs in the non-asymptotic data regime. |
Tasks | Bayesian Inference |
Published | 2018-09-06 |
URL | http://arxiv.org/abs/1809.02188v3 |
http://arxiv.org/pdf/1809.02188v3.pdf | |
PWC | https://paperswithcode.com/paper/differentially-private-bayesian-inference-for |
Repo | https://github.com/gbernstein6/private_bayesian_expfam |
Framework | none |
Evolving Multimodal Robot Behavior via Many Stepping Stones with the Combinatorial Multi-Objective Evolutionary Algorithm
Title | Evolving Multimodal Robot Behavior via Many Stepping Stones with the Combinatorial Multi-Objective Evolutionary Algorithm |
Authors | Joost Huizinga, Jeff Clune |
Abstract | An important challenge in reinforcement learning, including evolutionary robotics, is to solve multimodal problems, where agents have to act in qualitatively different ways depending on the circumstances. Because multimodal problems are often too difficult to solve directly, it is helpful to take advantage of staging, where a difficult task is divided into simpler subtasks that can serve as stepping stones for solving the overall problem. Unfortunately, choosing an effective ordering for these subtasks is difficult, and a poor ordering can reduce the speed and performance of the learning process. Here, we provide a thorough introduction and investigation of the Combinatorial Multi-Objective Evolutionary Algorithm (CMOEA), which avoids ordering subtasks by allowing all combinations of subtasks to be explored simultaneously. We compare CMOEA against two algorithms that can similarly optimize on multiple subtasks simultaneously: NSGA-II and Lexicase Selection. The algorithms are tested on a multimodal robotics problem with six subtasks as well as a maze navigation problem with a hundred subtasks. On these problems, CMOEA either outperforms or is competitive with the controls. Separately, we show that adding a linear combination over all objectives can improve the ability of NSGA-II to solve these multimodal problems. Lastly, we show that, in contrast to NSGA-II and Lexicase Selection, CMOEA can effectively leverage secondary objectives to achieve state-of-the-art results on the robotics task. In general, our experiments suggest that CMOEA is a promising, state-of-the-art algorithm for solving multimodal problems. |
Tasks | |
Published | 2018-07-09 |
URL | https://arxiv.org/abs/1807.03392v2 |
https://arxiv.org/pdf/1807.03392v2.pdf | |
PWC | https://paperswithcode.com/paper/evolving-multimodal-robot-behavior-via-many |
Repo | https://github.com/Evolving-AI-Lab/cmoea |
Framework | none |
Found Graph Data and Planted Vertex Covers
Title | Found Graph Data and Planted Vertex Covers |
Authors | Austin R. Benson, Jon Kleinberg |
Abstract | A typical way in which network data is recorded is to measure all the interactions among a specified set of core nodes; this produces a graph containing this core together with a potentially larger set of fringe nodes that have links to the core. Interactions between pairs of nodes in the fringe, however, are not recorded by this process, and hence not present in the resulting graph data. For example, a phone service provider may only have records of calls in which at least one of the participants is a customer; this can include calls between a customer and a non-customer, but not between pairs of non-customers. Knowledge of which nodes belong to the core is an important piece of metadata that is crucial for interpreting the network dataset. But in many cases, this metadata is not available, either because it has been lost due to difficulties in data provenance, or because the network consists of found data obtained in settings such as counter-surveillance. This leads to a natural algorithmic problem, namely the recovery of the core set. Since the core set forms a vertex cover of the graph, we essentially have a planted vertex cover problem, but with an arbitrary underlying graph. We develop a theoretical framework for analyzing this planted vertex cover problem, based on results in the theory of fixed-parameter tractability, together with algorithms for recovering the core. Our algorithms are fast, simple to implement, and out-perform several methods based on network core-periphery structure on various real-world datasets. |
Tasks | |
Published | 2018-05-03 |
URL | http://arxiv.org/abs/1805.01209v1 |
http://arxiv.org/pdf/1805.01209v1.pdf | |
PWC | https://paperswithcode.com/paper/found-graph-data-and-planted-vertex-covers |
Repo | https://github.com/arbenson/FGDnPVC |
Framework | none |
Deblending galaxy superpositions with branched generative adversarial networks
Title | Deblending galaxy superpositions with branched generative adversarial networks |
Authors | David M. Reiman, Brett E. Göhre |
Abstract | Near-future large galaxy surveys will encounter blended galaxy images at a fraction of up to 50% in the densest regions of the universe. Current deblending techniques may segment the foreground galaxy while leaving missing pixel intensities in the background galaxy flux. The problem is compounded by the diffuse nature of galaxies in their outer regions, making segmentation significantly more difficult than in traditional object segmentation applications. We propose a novel branched generative adversarial network (GAN) to deblend overlapping galaxies, where the two branches produce images of the two deblended galaxies. We show that generative models are a powerful engine for deblending given their innate ability to infill missing pixel values occluded by the superposition. We maintain high peak signal-to-noise ratio and structural similarity scores with respect to ground truth images upon deblending. Our model also predicts near-instantaneously, making it a natural choice for the immense quantities of data soon to be created by large surveys such as LSST, Euclid and WFIRST. |
Tasks | Semantic Segmentation |
Published | 2018-10-23 |
URL | http://arxiv.org/abs/1810.10098v4 |
http://arxiv.org/pdf/1810.10098v4.pdf | |
PWC | https://paperswithcode.com/paper/deblending-galaxy-superpositions-with |
Repo | https://github.com/davidreiman/deblender |
Framework | tf |
Scientific image rendering for space scenes with the SurRender software
Title | Scientific image rendering for space scenes with the SurRender software |
Authors | Roland Brochard, Jérémy Lebreton, Cyril Robin, Keyvan Kanani, Grégory Jonniaux, Aurore Masson, Noela Despré, Ahmad Berjaoui |
Abstract | Spacecraft autonomy can be enhanced by vision-based navigation (VBN) techniques. Applications range from manoeuvers around Solar System objects and landing on planetary surfaces, to in-orbit servicing or space debris removal. The development and validation of VBN algorithms relies on the availability of physically accurate relevant images. Yet archival data from past missions can rarely serve this purpose and acquiring new data is often costly. The SurRender software is an image simulator that addresses the challenges of realistic image rendering, with high representativeness for space scenes. Images are rendered by raytracing, which implements the physical principles of geometrical light propagation, in physical units. A macroscopic instrument model and scene objects reflectance functions are used. SurRender is specially optimized for space scenes, with huge distances between objects and scenes up to Solar System size. Raytracing conveniently tackles some important effects for VBN algorithms: image quality, eclipses, secondary illumination, subpixel limb imaging, etc. A simulation is easily setup (in MATLAB, Python, and more) by specifying the position of the bodies (camera, Sun, planets, satellites) over time, 3D shapes and material surface properties. SurRender comes with its own modelling tool enabling to go beyond existing models for shapes, materials and sensors (projection, temporal sampling, electronics, etc.). It is natively designed to simulate different kinds of sensors (visible, LIDAR, etc.). Tools are available for manipulating huge datasets to store albedo maps and digital elevation models, or for procedural (fractal) texturing that generates high-quality images for a large range of observing distances (from millions of km to touchdown). We illustrate SurRender performances with a selection of case studies, placing particular emphasis on a 900-km Moon flyby simulation. |
Tasks | |
Published | 2018-10-02 |
URL | http://arxiv.org/abs/1810.01423v1 |
http://arxiv.org/pdf/1810.01423v1.pdf | |
PWC | https://paperswithcode.com/paper/scientific-image-rendering-for-space-scenes |
Repo | https://github.com/SurRenderSoftware/surrender_client_API |
Framework | none |
Detection-Tracking for Efficient Person Analysis: The DetTA Pipeline
Title | Detection-Tracking for Efficient Person Analysis: The DetTA Pipeline |
Authors | Stefan Breuers, Lucas Beyer, Umer Rafi, Bastian Leibe |
Abstract | In the past decade many robots were deployed in the wild, and people detection and tracking is an important component of such deployments. On top of that, one often needs to run modules which analyze persons and extract higher level attributes such as age and gender, or dynamic information like gaze and pose. The latter ones are especially necessary for building a reactive, social robot-person interaction. In this paper, we combine those components in a fully modular detection-tracking-analysis pipeline, called DetTA. We investigate the benefits of such an integration on the example of head and skeleton pose, by using the consistent track ID for a temporal filtering of the analysis modules’ observations, showing a slight improvement in a challenging real-world scenario. We also study the potential of a so-called “free-flight” mode, where the analysis of a person attribute only relies on the filter’s predictions for certain frames. Here, our study shows that this boosts the runtime dramatically, while the prediction quality remains stable. This insight is especially important for reducing power consumption and sharing precious (GPU-)memory when running many analysis components on a mobile platform, especially so in the era of expensive deep learning methods. |
Tasks | |
Published | 2018-04-26 |
URL | http://arxiv.org/abs/1804.10134v2 |
http://arxiv.org/pdf/1804.10134v2.pdf | |
PWC | https://paperswithcode.com/paper/detection-tracking-for-efficient-person |
Repo | https://github.com/sbreuers/detta |
Framework | none |
LRW-1000: A Naturally-Distributed Large-Scale Benchmark for Lip Reading in the Wild
Title | LRW-1000: A Naturally-Distributed Large-Scale Benchmark for Lip Reading in the Wild |
Authors | Shuang Yang, Yuanhang Zhang, Dalu Feng, Mingmin Yang, Chenhao Wang, Jingyun Xiao, Keyu Long, Shiguang Shan, Xilin Chen |
Abstract | Large-scale datasets have successively proven their fundamental importance in several research fields, especially for early progress in some emerging topics. In this paper, we focus on the problem of visual speech recognition, also known as lipreading, which has received increasing interest in recent years. We present a naturally-distributed large-scale benchmark for lip reading in the wild, named LRW-1000, which contains 1,000 classes with 718,018 samples from more than 2,000 individual speakers. Each class corresponds to the syllables of a Mandarin word composed of one or several Chinese characters. To the best of our knowledge, it is currently the largest word-level lipreading dataset and also the only public large-scale Mandarin lipreading dataset. This dataset aims at covering a “natural” variability over different speech modes and imaging conditions to incorporate challenges encountered in practical applications. It has shown a large variation in this benchmark in several aspects, including the number of samples in each class, video resolution, lighting conditions, and speakers’ attributes such as pose, age, gender, and make-up. Besides providing a detailed description of the dataset and its collection pipeline, we evaluate several typical popular lipreading methods and perform a thorough analysis of the results from several aspects. The results demonstrate the consistency and challenges of our dataset, which may open up some new promising directions for future work. |
Tasks | Lipreading, Speech Recognition, Visual Speech Recognition |
Published | 2018-10-16 |
URL | http://arxiv.org/abs/1810.06990v6 |
http://arxiv.org/pdf/1810.06990v6.pdf | |
PWC | https://paperswithcode.com/paper/lrw-1000-a-naturally-distributed-large-scale |
Repo | https://github.com/Fengdalu/Lipreading-DenseNet3D |
Framework | pytorch |
Safe Option-Critic: Learning Safety in the Option-Critic Architecture
Title | Safe Option-Critic: Learning Safety in the Option-Critic Architecture |
Authors | Arushi Jain, Khimya Khetarpal, Doina Precup |
Abstract | Designing hierarchical reinforcement learning algorithms that induce a notion of safety is not only vital for safety-critical applications, but also, brings better understanding of an artificially intelligent agent’s decisions. While learning end-to-end options automatically has been fully realized recently, we propose a solution to learning safe options. We introduce the idea of controllability of states based on the temporal difference errors in the option-critic framework. We then derive the policy-gradient theorem with controllability and propose a novel framework called safe option-critic. We demonstrate the effectiveness of our approach in the four-rooms grid-world, cartpole, and three games in the Arcade Learning Environment (ALE): MsPacman, Amidar and Q*Bert. Learning of end-to-end options with the proposed notion of safety achieves reduction in the variance of return and boosts the performance in environments with intrinsic variability in the reward structure. More importantly, the proposed algorithm outperforms the vanilla options in all the environments and primitive actions in two out of three ALE games. |
Tasks | Atari Games, Hierarchical Reinforcement Learning |
Published | 2018-07-21 |
URL | http://arxiv.org/abs/1807.08060v1 |
http://arxiv.org/pdf/1807.08060v1.pdf | |
PWC | https://paperswithcode.com/paper/safe-option-critic-learning-safety-in-the |
Repo | https://github.com/arushi12130/SafeOptionCritic |
Framework | none |
Unsupervised adversarial domain adaptation for acoustic scene classification
Title | Unsupervised adversarial domain adaptation for acoustic scene classification |
Authors | Shayan Gharib, Konstantinos Drossos, Emre Çakir, Dmitriy Serdyuk, Tuomas Virtanen |
Abstract | A general problem in acoustic scene classification task is the mismatched conditions between training and testing data, which significantly reduces the performance of the developed methods on classification accuracy. As a countermeasure, we present the first method of unsupervised adversarial domain adaptation for acoustic scene classification. We employ a model pre-trained on data from one set of conditions and by using data from other set of conditions, we adapt the model in order that its output cannot be used for classifying the set of conditions that input data belong to. We use a freely available dataset from the DCASE 2018 challenge Task 1, subtask B, that contains data from mismatched recording devices. We consider the scenario where the annotations are available for the data recorded from one device, but not for the rest. Our results show that with our model agnostic method we can achieve $\sim 10%$ increase at the accuracy on an unseen and unlabeled dataset, while keeping almost the same performance on the labeled dataset. |
Tasks | Acoustic Scene Classification, Domain Adaptation, Scene Classification |
Published | 2018-08-17 |
URL | http://arxiv.org/abs/1808.05777v1 |
http://arxiv.org/pdf/1808.05777v1.pdf | |
PWC | https://paperswithcode.com/paper/unsupervised-adversarial-domain-adaptation |
Repo | https://github.com/shayangharib/AUDASC |
Framework | pytorch |
Stochastic Wasserstein Barycenters
Title | Stochastic Wasserstein Barycenters |
Authors | Sebastian Claici, Edward Chien, Justin Solomon |
Abstract | We present a stochastic algorithm to compute the barycenter of a set of probability distributions under the Wasserstein metric from optimal transport. Unlike previous approaches, our method extends to continuous input distributions and allows the support of the barycenter to be adjusted in each iteration. We tackle the problem without regularization, allowing us to recover a sharp output whose support is contained within the support of the true barycenter. We give examples where our algorithm recovers a more meaningful barycenter than previous work. Our method is versatile and can be extended to applications such as generating super samples from a given distribution and recovering blue noise approximations. |
Tasks | |
Published | 2018-02-15 |
URL | http://arxiv.org/abs/1802.05757v3 |
http://arxiv.org/pdf/1802.05757v3.pdf | |
PWC | https://paperswithcode.com/paper/stochastic-wasserstein-barycenters |
Repo | https://github.com/gparracl/StochasticWassersteinBarycenter |
Framework | tf |
PeerNets: Exploiting Peer Wisdom Against Adversarial Attacks
Title | PeerNets: Exploiting Peer Wisdom Against Adversarial Attacks |
Authors | Jan Svoboda, Jonathan Masci, Federico Monti, Michael M. Bronstein, Leonidas Guibas |
Abstract | Deep learning systems have become ubiquitous in many aspects of our lives. Unfortunately, it has been shown that such systems are vulnerable to adversarial attacks, making them prone to potential unlawful uses. Designing deep neural networks that are robust to adversarial attacks is a fundamental step in making such systems safer and deployable in a broader variety of applications (e.g. autonomous driving), but more importantly is a necessary step to design novel and more advanced architectures built on new computational paradigms rather than marginally building on the existing ones. In this paper we introduce PeerNets, a novel family of convolutional networks alternating classical Euclidean convolutions with graph convolutions to harness information from a graph of peer samples. This results in a form of non-local forward propagation in the model, where latent features are conditioned on the global structure induced by the graph, that is up to 3 times more robust to a variety of white- and black-box adversarial attacks compared to conventional architectures with almost no drop in accuracy. |
Tasks | Autonomous Driving |
Published | 2018-05-31 |
URL | http://arxiv.org/abs/1806.00088v1 |
http://arxiv.org/pdf/1806.00088v1.pdf | |
PWC | https://paperswithcode.com/paper/peernets-exploiting-peer-wisdom-against |
Repo | https://github.com/tantara/PeerNets-pytorch |
Framework | pytorch |
Deep Convolutional Networks as shallow Gaussian Processes
Title | Deep Convolutional Networks as shallow Gaussian Processes |
Authors | Adrià Garriga-Alonso, Carl Edward Rasmussen, Laurence Aitchison |
Abstract | We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior over the weights and biases is a Gaussian process (GP) in the limit of infinitely many convolutional filters, extending similar results for dense networks. For a CNN, the equivalent kernel can be computed exactly and, unlike “deep kernels”, has very few parameters: only the hyperparameters of the original CNN. Further, we show that this kernel has two properties that allow it to be computed efficiently; the cost of evaluating the kernel for a pair of images is similar to a single forward pass through the original CNN with only one filter per layer. The kernel equivalent to a 32-layer ResNet obtains 0.84% classification error on MNIST, a new record for GPs with a comparable number of parameters. |
Tasks | Gaussian Processes |
Published | 2018-08-16 |
URL | https://arxiv.org/abs/1808.05587v2 |
https://arxiv.org/pdf/1808.05587v2.pdf | |
PWC | https://paperswithcode.com/paper/deep-convolutional-networks-as-shallow |
Repo | https://github.com/convnets-as-gps/convnets-as-gps |
Framework | tf |
LadderNet: Multi-path networks based on U-Net for medical image segmentation
Title | LadderNet: Multi-path networks based on U-Net for medical image segmentation |
Authors | Juntang Zhuang |
Abstract | U-Net has been providing state-of-the-art performance in many medical image segmentation problems. Many modifications have been proposed for U-Net, such as attention U-Net, recurrent residual convolutional U-Net (R2-UNet), and U-Net with residual blocks or blocks with dense connections. However, all these modifications have an encoder-decoder structure with skip connections, and the number of paths for information flow is limited. We propose LadderNet in this paper, which can be viewed as a chain of multiple U-Nets. Instead of only one pair of encoder branch and decoder branch in U-Net, a LadderNet has multiple pairs of encoder-decoder branches, and has skip connections between every pair of adjacent decoder and decoder branches in each level. Inspired by the success of ResNet and R2-UNet, we use modified residual blocks where two convolutional layers in one block share the same weights. A LadderNet has more paths for information flow because of skip connections and residual blocks, and can be viewed as an ensemble of Fully Convolutional Networks (FCN). The equivalence to an ensemble of FCNs improves segmentation accuracy, while the shared weights within each residual block reduce parameter number. Semantic segmentation is essential for retinal disease detection. We tested LadderNet on two benchmark datasets for blood vessel segmentation in retinal images, and achieved superior performance over methods in the literature. The implementation is provided \url{https://github.com/juntang-zhuang/LadderNet} |
Tasks | Medical Image Segmentation, Retinal Vessel Segmentation, Semantic Segmentation |
Published | 2018-10-17 |
URL | https://arxiv.org/abs/1810.07810v4 |
https://arxiv.org/pdf/1810.07810v4.pdf | |
PWC | https://paperswithcode.com/paper/laddernet-multi-path-networks-based-on-u-net |
Repo | https://github.com/juntang-zhuang/LadderNet |
Framework | pytorch |
Stable Gaussian Process based Tracking Control of Euler-Lagrange Systems
Title | Stable Gaussian Process based Tracking Control of Euler-Lagrange Systems |
Authors | Thomas Beckers, Dana Kulić, Sandra Hirche |
Abstract | Perfect tracking control for real-world Euler-Lagrange systems is challenging due to uncertainties in the system model and external disturbances. The magnitude of the tracking error can be reduced either by increasing the feedback gains or improving the model of the system. The latter is clearly preferable as it allows to maintain good tracking performance at low feedback gains. However, accurate models are often difficult to obtain. In this article, we address the problem of stable high-performance tracking control for unknown Euler-Lagrange systems. In particular, we employ Gaussian Process regression to obtain a data-driven model that is used for the feed-forward compensation of unknown dynamics of the system. The model fidelity is used to adapt the feedback gains allowing low feedback gains in state space regions of high model confidence. The proposed control law guarantees a globally bounded tracking error with a specific probability. Simulation studies demonstrate the superiority over state of the art tracking control approaches. |
Tasks | |
Published | 2018-06-19 |
URL | http://arxiv.org/abs/1806.07190v2 |
http://arxiv.org/pdf/1806.07190v2.pdf | |
PWC | https://paperswithcode.com/paper/stable-gaussian-process-based-tracking |
Repo | https://github.com/TBeckers/CTC_GPR |
Framework | none |
Reconciling $λ$-Returns with Experience Replay
Title | Reconciling $λ$-Returns with Experience Replay |
Authors | Brett Daley, Christopher Amato |
Abstract | Modern deep reinforcement learning methods have departed from the incremental learning required for eligibility traces, rendering the implementation of the $\lambda$-return difficult in this context. In particular, off-policy methods that utilize experience replay remain problematic because their random sampling of minibatches is not conducive to the efficient calculation of $\lambda$-returns. Yet replay-based methods are often the most sample efficient, and incorporating $\lambda$-returns into them is a viable way to achieve new state-of-the-art performance. Towards this, we propose the first method to enable practical use of $\lambda$-returns in arbitrary replay-based methods without relying on other forms of decorrelation such as asynchronous gradient updates. By promoting short sequences of past transitions into a small cache within the replay memory, adjacent $\lambda$-returns can be efficiently precomputed by sharing Q-values. Computation is not wasted on experiences that are never sampled, and stored $\lambda$-returns behave as stable temporal-difference (TD) targets that replace the target network. Additionally, our method grants the unique ability to observe TD errors prior to sampling; for the first time, transitions can be prioritized by their true significance rather than by a proxy to it. Furthermore, we propose the novel use of the TD error to dynamically select $\lambda$-values that facilitate faster learning. We show that these innovations can enhance the performance of DQN when playing Atari 2600 games, even under partial observability. While our work specifically focuses on $\lambda$-returns, these ideas are applicable to any multi-step return estimator. |
Tasks | Atari Games |
Published | 2018-10-23 |
URL | https://arxiv.org/abs/1810.09967v3 |
https://arxiv.org/pdf/1810.09967v3.pdf | |
PWC | https://paperswithcode.com/paper/efficient-eligibility-traces-for-deep |
Repo | https://github.com/brett-daley/dqn-lambda |
Framework | tf |