Paper Group AWR 345
A Hybrid Compact Neural Architecture for Visual Place Recognition. Help, Anna! Visual Navigation with Natural Multimodal Assistance via Retrospective Curiosity-Encouraging Imitation Learning. Optimizing thermodynamic trajectories using evolutionary reinforcement learning. An Open Source and Open Hardware Deep Learning-powered Visual Navigation Engi …
A Hybrid Compact Neural Architecture for Visual Place Recognition
Title | A Hybrid Compact Neural Architecture for Visual Place Recognition |
Authors | Marvin Chancán, Luis Hernandez-Nunez, Ajay Narendra, Andrew B. Barron, Michael Milford |
Abstract | State-of-the-art algorithms for visual place recognition, and related visual navigation systems, can be broadly split into two categories: computer-science-oriented models including deep learning or image retrieval-based techniques with minimal biological plausibility, and neuroscience-oriented dynamical networks that model temporal properties underlying spatial navigation in the brain. In this letter, we propose a new compact and high-performing place recognition model that bridges this divide for the first time. Our approach comprises two key neural models of these categories: (1) FlyNet, a compact, sparse two-layer neural network inspired by brain architectures of fruit flies, Drosophila melanogaster, and (2) a one-dimensional continuous attractor neural network (CANN). The resulting FlyNet+CANN network incorporates the compact pattern recognition capabilities of our FlyNet model with the powerful temporal filtering capabilities of an equally compact CANN, replicating entirely in a hybrid neural implementation the functionality that yields high performance in algorithmic localization approaches like SeqSLAM. We evaluate our model, and compare it to three state-of-the-art methods, on two benchmark real-world datasets with small viewpoint variations and extreme environmental changes - achieving 87% AUC results under day to night transitions compared to 60% for Multi-Process Fusion, 46% for LoST-X and 1% for SeqSLAM, while being 6.5, 310, and 1.5 times faster, respectively. |
Tasks | Image Retrieval, Visual Navigation, Visual Place Recognition |
Published | 2019-10-15 |
URL | https://arxiv.org/abs/1910.06840v3 |
https://arxiv.org/pdf/1910.06840v3.pdf | |
PWC | https://paperswithcode.com/paper/a-compact-neural-architecture-for-visual |
Repo | https://github.com/mchancan/flynet |
Framework | pytorch |
Help, Anna! Visual Navigation with Natural Multimodal Assistance via Retrospective Curiosity-Encouraging Imitation Learning
Title | Help, Anna! Visual Navigation with Natural Multimodal Assistance via Retrospective Curiosity-Encouraging Imitation Learning |
Authors | Khanh Nguyen, Hal Daumé III |
Abstract | Mobile agents that can leverage help from humans can potentially accomplish more complex tasks than they could entirely on their own. We develop “Help, Anna!” (HANNA), an interactive photo-realistic simulator in which an agent fulfills object-finding tasks by requesting and interpreting natural language-and-vision assistance. An agent solving tasks in a HANNA environment can leverage simulated human assistants, called ANNA (Automatic Natural Navigation Assistants), which, upon request, provide natural language and visual instructions to direct the agent towards the goals. To address the HANNA problem, we develop a memory-augmented neural agent that hierarchically models multiple levels of decision-making, and an imitation learning algorithm that teaches the agent to avoid repeating past mistakes while simultaneously predicting its own chances of making future progress. Empirically, our approach is able to ask for help more effectively than competitive baselines and, thus, attains higher task success rate on both previously seen and previously unseen environments. We publicly release code and data at https://github.com/khanhptnk/hanna . A video demo is available at https://youtu.be/18P94aaaLKg . |
Tasks | Decision Making, Imitation Learning, Visual Navigation |
Published | 2019-09-04 |
URL | https://arxiv.org/abs/1909.01871v6 |
https://arxiv.org/pdf/1909.01871v6.pdf | |
PWC | https://paperswithcode.com/paper/help-anna-visual-navigation-with-natural |
Repo | https://github.com/khanhptnk/hanna |
Framework | pytorch |
Optimizing thermodynamic trajectories using evolutionary reinforcement learning
Title | Optimizing thermodynamic trajectories using evolutionary reinforcement learning |
Authors | Chris Beeler, Uladzimir Yahorau, Rory Coles, Kyle Mills, Stephen Whitelam, Isaac Tamblyn |
Abstract | Using a model heat engine we show that neural network-based reinforcement learning can identify thermodynamic trajectories of maximal efficiency. We use an evolutionary learning algorithm to evolve a population of neural networks, subject to a directive to maximize the efficiency of a trajectory composed of a set of elementary thermodynamic processes; the resulting networks learn to carry out the maximally-efficient Carnot, Stirling, or Otto cycles. Given additional irreversible processes this evolutionary scheme learns a hitherto unknown thermodynamic cycle. Our results show how the reinforcement learning strategies developed for game playing can be applied to solve physical problems conditioned upon path-extensive order parameters. |
Tasks | |
Published | 2019-03-20 |
URL | http://arxiv.org/abs/1903.08543v1 |
http://arxiv.org/pdf/1903.08543v1.pdf | |
PWC | https://paperswithcode.com/paper/optimizing-thermodynamic-trajectories-using |
Repo | https://github.com/CLEANit/heatenginegym |
Framework | none |
An Open Source and Open Hardware Deep Learning-powered Visual Navigation Engine for Autonomous Nano-UAVs
Title | An Open Source and Open Hardware Deep Learning-powered Visual Navigation Engine for Autonomous Nano-UAVs |
Authors | Daniele Palossi, Francesco Conti, Luca Benini |
Abstract | Nano-size unmanned aerial vehicles (UAVs), with few centimeters of diameter and sub-10 Watts of total power budget, have so far been considered incapable of running sophisticated visual-based autonomous navigation software without external aid from base-stations, ad-hoc local positioning infrastructure, and powerful external computation servers. In this work, we present what is, to the best of our knowledge, the first 27g nano-UAV system able to run aboard an end-to-end, closed-loop visual pipeline for autonomous navigation based on a state-of-the-art deep-learning algorithm, built upon the open-source CrazyFlie 2.0 nano-quadrotor. Our visual navigation engine is enabled by the combination of an ultra-low power computing device (the GAP8 system-on-chip) with a novel methodology for the deployment of deep convolutional neural networks (CNNs). We enable onboard real-time execution of a state-of-the-art deep CNN at up to 18Hz. Field experiments demonstrate that the system’s high responsiveness prevents collisions with unexpected dynamic obstacles up to a flight speed of 1.5m/s. In addition, we also demonstrate the capability of our visual navigation engine of fully autonomous indoor navigation on a 113m previously unseen path. To share our key findings with the embedded and robotics communities and foster further developments in autonomous nano-UAVs, we publicly release all our code, datasets, and trained networks. |
Tasks | Autonomous Navigation, Visual Navigation |
Published | 2019-05-10 |
URL | https://arxiv.org/abs/1905.04166v1 |
https://arxiv.org/pdf/1905.04166v1.pdf | |
PWC | https://paperswithcode.com/paper/an-open-source-and-open-hardware-deep |
Repo | https://github.com/pulp-platform/pulp-dronet |
Framework | none |
hyppo: A Comprehensive Multivariate Hypothesis Testing Python Package
Title | hyppo: A Comprehensive Multivariate Hypothesis Testing Python Package |
Authors | Sambit Panda, Satish Palaniappan, Junhao Xiong, Eric W. Bridgeford, Ronak Mehta, Cencheng Shen, Joshua T. Vogelstein |
Abstract | We introduce hyppo, a unified library for performing multivariate hypothesis testing, including independence, two-sample, and k-sample testing. While many multivariate independence tests have R packages available, the interfaces are inconsistent and most are not available in Python. hyppo includes many state of the art multivariate testing procedures. The package is easy-to-use and is flexible enough to enable future extensions. The documentation and all releases are available at https://hyppo.neurodata.io. |
Tasks | |
Published | 2019-07-03 |
URL | https://arxiv.org/abs/1907.02088v3 |
https://arxiv.org/pdf/1907.02088v3.pdf | |
PWC | https://paperswithcode.com/paper/mgcpy-a-comprehensive-high-dimensional |
Repo | https://github.com/neurodata/MGC |
Framework | none |
GraphACT: Accelerating GCN Training on CPU-FPGA Heterogeneous Platforms
Title | GraphACT: Accelerating GCN Training on CPU-FPGA Heterogeneous Platforms |
Authors | Hanqing Zeng, Viktor Prasanna |
Abstract | Graph Convolutional Networks (GCNs) have emerged as the state-of-the-art deep learning model for representation learning on graphs. It is challenging to accelerate training of GCNs, due to (1) substantial and irregular data communication to propagate information within the graph, and (2) intensive computation to propagate information along the neural network layers. To address these challenges, we design a novel accelerator for training GCNs on CPU-FPGA heterogeneous systems, by incorporating multiple algorithm-architecture co-optimizations. We first analyze the computation and communication characteristics of various GCN training algorithms, and select a subgraph-based algorithm that is well suited for hardware execution. To optimize the feature propagation within subgraphs, we propose a lightweight pre-processing step based on a graph theoretic approach. Such pre-processing performed on the CPU significantly reduces the memory access requirements and the computation to be performed on the FPGA. To accelerate the weight update in GCN layers, we propose a systolic array based design for efficient parallelization. We integrate the above optimizations into a complete hardware pipeline, and analyze its load-balance and resource utilization by accurate performance modeling. We evaluate our design on a Xilinx Alveo U200 board hosted by a 40-core Xeon server. On three large graphs, we achieve an order of magnitude training speedup with negligible accuracy loss, compared with state-of-the-art implementation on a multi-core platform. |
Tasks | Representation Learning |
Published | 2019-12-31 |
URL | https://arxiv.org/abs/2001.02498v1 |
https://arxiv.org/pdf/2001.02498v1.pdf | |
PWC | https://paperswithcode.com/paper/graphact-accelerating-gcn-training-on-cpu |
Repo | https://github.com/GraphSAINT/GraphACT |
Framework | none |
Centripetal SGD for Pruning Very Deep Convolutional Networks with Complicated Structure
Title | Centripetal SGD for Pruning Very Deep Convolutional Networks with Complicated Structure |
Authors | Xiaohan Ding, Guiguang Ding, Yuchen Guo, Jungong Han |
Abstract | The redundancy is widely recognized in Convolutional Neural Networks (CNNs), which enables to remove unimportant filters from convolutional layers so as to slim the network with acceptable performance drop. Inspired by the linear and combinational properties of convolution, we seek to make some filters increasingly close and eventually identical for network slimming. To this end, we propose Centripetal SGD (C-SGD), a novel optimization method, which can train several filters to collapse into a single point in the parameter hyperspace. When the training is completed, the removal of the identical filters can trim the network with NO performance loss, thus no finetuning is needed. By doing so, we have partly solved an open problem of constrained filter pruning on CNNs with complicated structure, where some layers must be pruned following others. Our experimental results on CIFAR-10 and ImageNet have justified the effectiveness of C-SGD-based filter pruning. Moreover, we have provided empirical evidences for the assumption that the redundancy in deep neural networks helps the convergence of training by showing that a redundant CNN trained using C-SGD outperforms a normally trained counterpart with the equivalent width. |
Tasks | |
Published | 2019-04-08 |
URL | http://arxiv.org/abs/1904.03837v1 |
http://arxiv.org/pdf/1904.03837v1.pdf | |
PWC | https://paperswithcode.com/paper/centripetal-sgd-for-pruning-very-deep |
Repo | https://github.com/ShawnDing1994/Centripetal-SGD |
Framework | tf |
Anatomical Priors in Convolutional Networks for Unsupervised Biomedical Segmentation
Title | Anatomical Priors in Convolutional Networks for Unsupervised Biomedical Segmentation |
Authors | Adrian V. Dalca, John Guttag, Mert R. Sabuncu |
Abstract | We consider the problem of segmenting a biomedical image into anatomical regions of interest. We specifically address the frequent scenario where we have no paired training data that contains images and their manual segmentations. Instead, we employ unpaired segmentation images to build an anatomical prior. Critically these segmentations can be derived from imaging data from a different dataset and imaging modality than the current task. We introduce a generative probabilistic model that employs the learned prior through a convolutional neural network to compute segmentations in an unsupervised setting. We conducted an empirical analysis of the proposed approach in the context of structural brain MRI segmentation, using a multi-study dataset of more than 14,000 scans. Our results show that an anatomical prior can enable fast unsupervised segmentation which is typically not possible using standard convolutional networks. The integration of anatomical priors can facilitate CNN-based anatomical segmentation in a range of novel clinical problems, where few or no annotations are available and thus standard networks are not trainable. The code is freely available at http://github.com/adalca/neuron. |
Tasks | |
Published | 2019-03-07 |
URL | http://arxiv.org/abs/1903.03148v1 |
http://arxiv.org/pdf/1903.03148v1.pdf | |
PWC | https://paperswithcode.com/paper/anatomical-priors-in-convolutional-networks-1 |
Repo | https://github.com/adalca/neuron |
Framework | tf |
Compositional generalization through meta sequence-to-sequence learning
Title | Compositional generalization through meta sequence-to-sequence learning |
Authors | Brenden M. Lake |
Abstract | People can learn a new concept and use it compositionally, understanding how to “blicket twice” after learning how to “blicket.” In contrast, powerful sequence-to-sequence (seq2seq) neural networks fail such tests of compositionality, especially when composing new concepts together with existing concepts. In this paper, I show how memory-augmented neural networks can be trained to generalize compositionally through meta seq2seq learning. In this approach, models train on a series of seq2seq problems to acquire the compositional skills needed to solve new seq2seq problems. Meta se2seq learning solves several of the SCAN tests for compositional learning and can learn to apply implicit rules to variables. |
Tasks | |
Published | 2019-06-12 |
URL | https://arxiv.org/abs/1906.05381v2 |
https://arxiv.org/pdf/1906.05381v2.pdf | |
PWC | https://paperswithcode.com/paper/compositional-generalization-through-meta |
Repo | https://github.com/brendenlake/meta_seq2seq |
Framework | pytorch |
Solving Interpretable Kernel Dimension Reduction
Title | Solving Interpretable Kernel Dimension Reduction |
Authors | Chieh Wu, Jared Miller, Yale Chang, Mario Sznaier, Jennifer Dy |
Abstract | Kernel dimensionality reduction (KDR) algorithms find a low dimensional representation of the original data by optimizing kernel dependency measures that are capable of capturing nonlinear relationships. The standard strategy is to first map the data into a high dimensional feature space using kernels prior to a projection onto a low dimensional space. While KDR methods can be easily solved by keeping the most dominant eigenvectors of the kernel matrix, its features are no longer easy to interpret. Alternatively, Interpretable KDR (IKDR) is different in that it projects onto a subspace \textit{before} the kernel feature mapping, therefore, the projection matrix can indicate how the original features linearly combine to form the new features. Unfortunately, the IKDR objective requires a non-convex manifold optimization that is difficult to solve and can no longer be solved by eigendecomposition. Recently, an efficient iterative spectral (eigendecomposition) method (ISM) has been proposed for this objective in the context of alternative clustering. However, ISM only provides theoretical guarantees for the Gaussian kernel. This greatly constrains ISM’s usage since any kernel method using ISM is now limited to a single kernel. This work extends the theoretical guarantees of ISM to an entire family of kernels, thereby empowering ISM to solve any kernel method of the same objective. In identifying this family, we prove that each kernel within the family has a surrogate $\Phi$ matrix and the optimal projection is formed by its most dominant eigenvectors. With this extension, we establish how a wide range of IKDR applications across different learning paradigms can be solved by ISM. To support reproducible results, the source code is made publicly available on \url{https://github.com/chieh-neu/ISM_supervised_DR}. |
Tasks | Dimensionality Reduction |
Published | 2019-09-06 |
URL | https://arxiv.org/abs/1909.03093v3 |
https://arxiv.org/pdf/1909.03093v3.pdf | |
PWC | https://paperswithcode.com/paper/solving-interpretable-kernel-dimension |
Repo | https://github.com/chieh-neu/ISM_supervised_DR |
Framework | none |
Beyond Human-Level Accuracy: Computational Challenges in Deep Learning
Title | Beyond Human-Level Accuracy: Computational Challenges in Deep Learning |
Authors | Joel Hestness, Newsha Ardalani, Greg Diamos |
Abstract | Deep learning (DL) research yields accuracy and product improvements from both model architecture changes and scale: larger data sets and models, and more computation. For hardware design, it is difficult to predict DL model changes. However, recent prior work shows that as dataset sizes grow, DL model accuracy and model size grow predictably. This paper leverages the prior work to project the dataset and model size growth required to advance DL accuracy beyond human-level, to frontier targets defined by machine learning experts. Datasets will need to grow $33$–$971 \times$, while models will need to grow $6.6$–$456\times$ to achieve target accuracies. We further characterize and project the computational requirements to train these applications at scale. Our characterization reveals an important segmentation of DL training challenges for recurrent neural networks (RNNs) that contrasts with prior studies of deep convolutional networks. RNNs will have comparatively moderate operational intensities and very large memory footprint requirements. In contrast to emerging accelerator designs, large-scale RNN training characteristics suggest designs with significantly larger memory capacity and on-chip caches. |
Tasks | |
Published | 2019-09-03 |
URL | https://arxiv.org/abs/1909.01736v1 |
https://arxiv.org/pdf/1909.01736v1.pdf | |
PWC | https://paperswithcode.com/paper/beyond-human-level-accuracy-computational |
Repo | https://github.com/baidu-research/catamount |
Framework | tf |
HyperNOMAD: Hyperparameter optimization of deep neural networks using mesh adaptive direct search
Title | HyperNOMAD: Hyperparameter optimization of deep neural networks using mesh adaptive direct search |
Authors | Dounia Lakhmiri, Sébastien Le Digabel, Christophe Tribes |
Abstract | The performance of deep neural networks is highly sensitive to the choice of the hyperparameters that define the structure of the network and the learning process. When facing a new application, tuning a deep neural network is a tedious and time consuming process that is often described as a “dark art”. This explains the necessity of automating the calibration of these hyperparameters. Derivative-free optimization is a field that develops methods designed to optimize time consuming functions without relying on derivatives. This work introduces the HyperNOMAD package, an extension of the NOMAD software that applies the MADS algorithm [7] to simultaneously tune the hyperparameters responsible for both the architecture and the learning process of a deep neural network (DNN), and that allows for an important flexibility in the exploration of the search space by taking advantage of categorical variables. This new approach is tested on the MNIST and CIFAR-10 data sets and achieves results comparable to the current state of the art. |
Tasks | Calibration, Hyperparameter Optimization |
Published | 2019-07-03 |
URL | https://arxiv.org/abs/1907.01698v1 |
https://arxiv.org/pdf/1907.01698v1.pdf | |
PWC | https://paperswithcode.com/paper/hypernomad-hyperparameter-optimization-of |
Repo | https://github.com/DouniaLakhmiri/HyperNOMAD |
Framework | pytorch |
GFF: Gated Fully Fusion for Semantic Segmentation
Title | GFF: Gated Fully Fusion for Semantic Segmentation |
Authors | Xiangtai Li, Houlong Zhao, Lei Han, Yunhai Tong, Kuiyuan Yang |
Abstract | Semantic segmentation generates comprehensive understanding of scenes through densely predicting the category for each pixel. High-level features from Deep Convolutional Neural Networks already demonstrate their effectiveness in semantic segmentation tasks, however the coarse resolution of high-level features often leads to inferior results for small/thin objects where detailed information is important. It is natural to consider importing low level features to compensate for the lost detailed information in high-level features.Unfortunately, simply combining multi-level features suffers from the semantic gap among them. In this paper, we propose a new architecture, named Gated Fully Fusion (GFF), to selectively fuse features from multiple levels using gates in a fully connected way. Specifically, features at each level are enhanced by higher-level features with stronger semantics and lower-level features with more details, and gates are used to control the propagation of useful information which significantly reduces the noises during fusion. We achieve the state of the art results on four challenging scene parsing datasets including Cityscapes, Pascal Context, COCO-stuff and ADE20K. |
Tasks | Scene Parsing, Scene Understanding, Semantic Segmentation |
Published | 2019-04-03 |
URL | https://arxiv.org/abs/1904.01803v2 |
https://arxiv.org/pdf/1904.01803v2.pdf | |
PWC | https://paperswithcode.com/paper/gff-gated-fully-fusion-for-semantic |
Repo | https://github.com/nhatuan84/GFF-Gated-Fully-Fusion-for-Semantic-Segmentation |
Framework | tf |
Harmonic Networks with Limited Training Samples
Title | Harmonic Networks with Limited Training Samples |
Authors | Matej Ulicny, Vladimir A. Krylov, Rozenn Dahyot |
Abstract | Convolutional neural networks (CNNs) are very popular nowadays for image processing. CNNs allow one to learn optimal filters in a (mostly) supervised machine learning context. However this typically requires abundant labelled training data to estimate the filter parameters. Alternative strategies have been deployed for reducing the number of parameters and / or filters to be learned and thus decrease overfitting. In the context of reverting to preset filters, we propose here a computationally efficient harmonic block that uses Discrete Cosine Transform (DCT) filters in CNNs. In this work we examine the performance of harmonic networks in limited training data scenario. We validate experimentally that its performance compares well against scattering networks that use wavelets as preset filters. |
Tasks | Image Classification |
Published | 2019-04-30 |
URL | http://arxiv.org/abs/1905.00135v1 |
http://arxiv.org/pdf/1905.00135v1.pdf | |
PWC | https://paperswithcode.com/paper/harmonic-networks-with-limited-training |
Repo | https://github.com/matej-ulicny/harmonic-networks |
Framework | pytorch |
Extractive Summarization of Long Documents by Combining Global and Local Context
Title | Extractive Summarization of Long Documents by Combining Global and Local Context |
Authors | Wen Xiao, Giuseppe Carenini |
Abstract | In this paper, we propose a novel neural single document extractive summarization model for long documents, incorporating both the global context of the whole document and the local context within the current topic. We evaluate the model on two datasets of scientific papers, Pubmed and arXiv, where it outperforms previous work, both extractive and abstractive models, on ROUGE-1, ROUGE-2 and METEOR scores. We also show that, consistently with our goal, the benefits of our method become stronger as we apply it to longer documents. Rather surprisingly, an ablation study indicates that the benefits of our model seem to come exclusively from modeling the local context, even for the longest documents. |
Tasks | |
Published | 2019-09-17 |
URL | https://arxiv.org/abs/1909.08089v1 |
https://arxiv.org/pdf/1909.08089v1.pdf | |
PWC | https://paperswithcode.com/paper/extractive-summarization-of-long-documents-by |
Repo | https://github.com/Wendy-Xiao/Extsumm_local_global_context |
Framework | pytorch |