Paper Group ANR 87
Understanding and Mitigating the Tradeoff Between Robustness and Accuracy. Axial-DeepLab: Stand-Alone Axial-Attention for Panoptic Segmentation. Deep Learning for Sensor-based Human Activity Recognition: Overview, Challenges and Opportunities. Unifying Training and Inference for Panoptic Segmentation. Evolution of Image Segmentation using Deep Conv …
Understanding and Mitigating the Tradeoff Between Robustness and Accuracy
Title | Understanding and Mitigating the Tradeoff Between Robustness and Accuracy |
Authors | Aditi Raghunathan, Sang Michael Xie, Fanny Yang, John Duchi, Percy Liang |
Abstract | Adversarial training augments the training set with perturbations to improve the robust error (over worst-case perturbations), but it often leads to an increase in the standard error (on unperturbed test inputs). Previous explanations for this tradeoff rely on the assumption that no predictor in the hypothesis class has low standard and robust error. In this work, we precisely characterize the effect of augmentation on the standard error in linear regression when the optimal linear predictor has zero standard and robust error. In particular, we show that the standard error could increase even when the augmented perturbations have noiseless observations from the optimal linear predictor. We then prove that the recently proposed robust self-training (RST) estimator improves robust error without sacrificing standard error for noiseless linear regression. Empirically, for neural networks, we find that RST with different adversarial training methods improves both standard and robust error for random and adversarial rotations and adversarial $\ell_\infty$ perturbations in CIFAR-10. |
Tasks | |
Published | 2020-02-25 |
URL | https://arxiv.org/abs/2002.10716v1 |
https://arxiv.org/pdf/2002.10716v1.pdf | |
PWC | https://paperswithcode.com/paper/understanding-and-mitigating-the-tradeoff |
Repo | |
Framework | |
Axial-DeepLab: Stand-Alone Axial-Attention for Panoptic Segmentation
Title | Axial-DeepLab: Stand-Alone Axial-Attention for Panoptic Segmentation |
Authors | Huiyu Wang, Yukun Zhu, Bradley Green, Hartwig Adam, Alan Yuille, Liang-Chieh Chen |
Abstract | Convolution exploits locality for efficiency at a cost of missing long range context. Self-attention has been adopted to augment CNNs with non-local interactions. Recent works prove it possible to stack self-attention layers to obtain a fully attentional network by restricting the attention to a local region. In this paper, we attempt to remove this constraint by factorizing 2D self-attention into two 1D self-attentions. This reduces computation complexity and allows performing attention within a larger or even global region. In companion, we also propose a position-sensitive self-attention design. Combining both yields our position-sensitive axial-attention layer, a novel building block that one could stack to form axial-attention models for image classification and dense prediction. We demonstrate the effectiveness of our model on four large-scale datasets. In particular, our model outperforms all existing stand-alone self-attention models on ImageNet. Our Axial-DeepLab improves 2.8% PQ over bottom-up state-of-the-art on COCO test-dev. This previous state-of-the-art is attained by our small variant that is 3.8x parameter-efficient and 27x computation-efficient. Axial-DeepLab also achieves state-of-the-art results on Mapillary Vistas and Cityscapes. |
Tasks | Image Classification, Panoptic Segmentation |
Published | 2020-03-17 |
URL | https://arxiv.org/abs/2003.07853v1 |
https://arxiv.org/pdf/2003.07853v1.pdf | |
PWC | https://paperswithcode.com/paper/axial-deeplab-stand-alone-axial-attention-for |
Repo | |
Framework | |
Deep Learning for Sensor-based Human Activity Recognition: Overview, Challenges and Opportunities
Title | Deep Learning for Sensor-based Human Activity Recognition: Overview, Challenges and Opportunities |
Authors | Kaixuan Chen, Dalin Zhang, Lina Yao, Bin Guo, Zhiwen Yu, Yunhao Liu |
Abstract | The vast proliferation of sensor devices and Internet of Things enables the applications of sensor-based activity recognition. However, there exist substantial challenges that could influence the performance of the recognition system in practical scenarios. Recently, as deep learning has demonstrated its effectiveness in many areas, plenty of deep methods have been investigated to address the challenges in activity recognition. In this study, we present a survey of the state-of-the-art deep learning methods for sensor-based human activity recognition. We first introduce the multi-modality of the sensory data and provide information for public datasets that can be used for evaluation in different challenge tasks. We then propose a new taxonomy to structure the deep methods by challenges. Challenges and challenge-related deep methods are summarized and analyzed to form an overview of the current research progress. At the end of this work, we discuss the open issues and provide some insights for future directions. |
Tasks | Activity Recognition, Human Activity Recognition |
Published | 2020-01-21 |
URL | https://arxiv.org/abs/2001.07416v1 |
https://arxiv.org/pdf/2001.07416v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-learning-for-sensor-based-human-activity |
Repo | |
Framework | |
Unifying Training and Inference for Panoptic Segmentation
Title | Unifying Training and Inference for Panoptic Segmentation |
Authors | Qizhu Li, Xiaojuan Qi, Philip H. S. Torr |
Abstract | We present an end-to-end network to bridge the gap between training and inference pipeline for panoptic segmentation, a task that seeks to partition an image into semantic regions for “stuff” and object instances for “things”. In contrast to recent works, our network exploits a parametrised, yet lightweight panoptic segmentation submodule, powered by an end-to-end learnt dense instance affinity, to capture the probability that any pair of pixels belong to the same instance. This panoptic submodule gives rise to a novel propagation mechanism for panoptic logits and enables the network to output a coherent panoptic segmentation map for both “stuff” and “thing” classes, without any post-processing. Reaping the benefits of end-to-end training, our full system sets new records on the popular street scene dataset, Cityscapes, achieving 61.4 PQ with a ResNet-50 backbone using only the fine annotations. On the challenging COCO dataset, our ResNet-50-based network also delivers state-of-the-art accuracy of 43.4 PQ. Moreover, our network flexibly works with and without object mask cues, performing competitively under both settings, which is of interest for applications with computation budgets. |
Tasks | Panoptic Segmentation |
Published | 2020-01-14 |
URL | https://arxiv.org/abs/2001.04982v1 |
https://arxiv.org/pdf/2001.04982v1.pdf | |
PWC | https://paperswithcode.com/paper/unifying-training-and-inference-for-panoptic |
Repo | |
Framework | |
Evolution of Image Segmentation using Deep Convolutional Neural Network: A Survey
Title | Evolution of Image Segmentation using Deep Convolutional Neural Network: A Survey |
Authors | Farhana Sultana, Abu Sufian, Paramartha Dutta |
Abstract | From the autonomous car driving to medical diagnosis, the requirement of the task of image segmentation is everywhere. Segmentation of an image is one of the indispensable tasks in computer vision. This task is comparatively complicated than other vision tasks as it needs low-level spatial information. Basically, image segmentation can be of two types: semantic segmentation and instance segmentation. The combined version of these two basic tasks is known as panoptic segmentation. In the recent era, the success of deep convolutional neural network (CNN) has influenced the field of segmentation greatly and gave us various successful models to date. In this survey, we are going to take a glance at the evolution of both semantic and instance segmentation work based on CNN. We have also specified comparative architectural details of some state-of-the-art models and discuss their training details to present a lucid understanding of hyper-parameter tuning of those models. Lastly, we have drawn a comparison among the performance of those models on different datasets. |
Tasks | Instance Segmentation, Medical Diagnosis, Panoptic Segmentation, Semantic Segmentation |
Published | 2020-01-13 |
URL | https://arxiv.org/abs/2001.04074v2 |
https://arxiv.org/pdf/2001.04074v2.pdf | |
PWC | https://paperswithcode.com/paper/evolution-of-image-segmentation-using-deep |
Repo | |
Framework | |
Effective End-to-End Learning Framework for Economic Dispatch
Title | Effective End-to-End Learning Framework for Economic Dispatch |
Authors | Chenbei Lu, Kui Wang, Chenye Wu |
Abstract | Conventional wisdom to improve the effectiveness of economic dispatch is to design the load forecasting method as accurately as possible. However, this approach can be problematic due to the temporal and spatial correlations between system cost and load prediction errors. This motivates us to adopt the notion of end-to-end machine learning and to propose a task-specific learning criteria to conduct economic dispatch. Specifically, to maximize the data utilization, we design an efficient optimization kernel for the learning process. We provide both theoretical analysis and empirical insights to highlight the effectiveness and efficiency of the proposed learning framework. |
Tasks | Load Forecasting |
Published | 2020-02-22 |
URL | https://arxiv.org/abs/2002.12755v1 |
https://arxiv.org/pdf/2002.12755v1.pdf | |
PWC | https://paperswithcode.com/paper/effective-end-to-end-learning-framework-for |
Repo | |
Framework | |
Forming Diverse Teams from Sequentially Arriving People
Title | Forming Diverse Teams from Sequentially Arriving People |
Authors | Faez Ahmed, John Dickerson, Mark Fuge |
Abstract | Collaborative work often benefits from having teams or organizations with heterogeneous members. In this paper, we present a method to form such diverse teams from people arriving sequentially over time. We define a monotone submodular objective function that combines the diversity and quality of a team and propose an algorithm to maximize the objective while satisfying multiple constraints. This allows us to balance both how diverse the team is and how well it can perform the task at hand. Using crowd experiments, we show that, in practice, the algorithm leads to large gains in team diversity. Using simulations, we show how to quantify the additional cost of forming diverse teams and how to address the problem of simultaneously maximizing diversity for several attributes (e.g., country of origin, gender). Our method has applications in collaborative work ranging from team formation, the assignment of workers to teams in crowdsourcing, and reviewer allocation to journal papers arriving sequentially. Our code is publicly accessible for further research. |
Tasks | |
Published | 2020-02-25 |
URL | https://arxiv.org/abs/2002.10697v1 |
https://arxiv.org/pdf/2002.10697v1.pdf | |
PWC | https://paperswithcode.com/paper/forming-diverse-teams-from-sequentially |
Repo | |
Framework | |
Efficient Tensor Kernel methods for sparse regression
Title | Efficient Tensor Kernel methods for sparse regression |
Authors | Feliks Hibraj, Marcello Pelillo, Saverio Salzo, Massimiliano Pontil |
Abstract | Recently, classical kernel methods have been extended by the introduction of suitable tensor kernels so to promote sparsity in the solution of the underlying regression problem. Indeed, they solve an lp-norm regularization problem, with p=m/(m-1) and m even integer, which happens to be close to a lasso problem. However, a major drawback of the method is that storing tensors requires a considerable amount of memory, ultimately limiting its applicability. In this work we address this problem by proposing two advances. First, we directly reduce the memory requirement, by intriducing a new and more efficient layout for storing the data. Second, we use a Nystrom-type subsampling approach, which allows for a training phase with a smaller number of data points, so to reduce the computational cost. Experiments, both on synthetic and read datasets, show the effectiveness of the proposed improvements. Finally, we take case of implementing the cose in C++ so to further speed-up the computation. |
Tasks | |
Published | 2020-03-23 |
URL | https://arxiv.org/abs/2003.10482v1 |
https://arxiv.org/pdf/2003.10482v1.pdf | |
PWC | https://paperswithcode.com/paper/efficient-tensor-kernel-methods-for-sparse |
Repo | |
Framework | |
Cooperation without Coordination: Hierarchical Predictive Planning for Decentralized Multiagent Navigation
Title | Cooperation without Coordination: Hierarchical Predictive Planning for Decentralized Multiagent Navigation |
Authors | Rose E. Wang, J. Chase Kew, Dennis Lee, Tsang-Wei Edward Lee, Tingnan Zhang, Brian Ichter, Jie Tan, Aleksandra Faust |
Abstract | Decentralized multiagent planning raises many challenges, such as adaption to changing environments inexplicable by the agent’s own behavior, coordination from noisy sensor inputs like lidar, cooperation without knowing other agents’ intents. To address these challenges, we present hierarchical predictive planning (HPP) for decentralized multiagent navigation tasks. HPP learns prediction models for itself and other teammates, and uses the prediction models to propose and evaluate navigation goals that complete the cooperative task without explicit coordination. To learn the prediction models, HPP observes other agents’ behavior and learns to maps own sensors to predicted locations of other agents. HPP then uses the cross-entropy method to iteratively propose, evaluate, and improve navigation goals, under assumption that all agents in the team share a common objective. HPP removes the need for a centralized operator (i.e. robots determine their own actions without coordinating their beliefs or plans) and can be trained and easily transferred to real world environments. The results show that HPP generalizes to new environments including real-world robot team. It is also 33x more sample efficient and performs better in complex environments compared to a baseline. The video and website for this paper can be found at https://youtu.be/-LqgfksqNH8 and https://sites.google.com/view/multiagent-hpp. |
Tasks | |
Published | 2020-03-15 |
URL | https://arxiv.org/abs/2003.06906v1 |
https://arxiv.org/pdf/2003.06906v1.pdf | |
PWC | https://paperswithcode.com/paper/cooperation-without-coordination-hierarchical |
Repo | |
Framework | |
Multi-task Learning for Voice Trigger Detection
Title | Multi-task Learning for Voice Trigger Detection |
Authors | Siddharth Sigtia, Pascal Clark, Rob Haynes, Hywel Richards, John Bridle |
Abstract | We describe the design of a voice trigger detection system for smart speakers. In this study, we address two major challenges. The first is that the detectors are deployed in complex acoustic environments with external noise and loud playback by the device itself. Secondly, collecting training examples for a specific keyword or trigger phrase is challenging resulting in a scarcity of trigger phrase specific training data. We describe a two-stage cascaded architecture where a low-power detector is always running and listening for the trigger phrase. If a detection is made at this stage, the candidate audio segment is re-scored by larger, more complex models to verify that the segment contains the trigger phrase. In this study, we focus our attention on the architecture and design of these second-pass detectors. We start by training a general acoustic model that produces phonetic transcriptions given a large labelled training dataset. Next, we collect a much smaller dataset of examples that are challenging for the baseline system. We then use multi-task learning to train a model to simultaneously produce accurate phonetic transcriptions on the larger dataset \emph{and} discriminate between true and easily confusable examples using the smaller dataset. Our results demonstrate that the proposed model reduces errors by half compared to the baseline in a range of challenging test conditions \emph{without} requiring extra parameters. |
Tasks | Multi-Task Learning |
Published | 2020-01-26 |
URL | https://arxiv.org/abs/2001.09519v1 |
https://arxiv.org/pdf/2001.09519v1.pdf | |
PWC | https://paperswithcode.com/paper/multi-task-learning-for-voice-trigger |
Repo | |
Framework | |
MT-BioNER: Multi-task Learning for Biomedical Named Entity Recognition using Deep Bidirectional Transformers
Title | MT-BioNER: Multi-task Learning for Biomedical Named Entity Recognition using Deep Bidirectional Transformers |
Authors | Muhammad Raza Khan, Morteza Ziyadi, Mohamed AbdelHady |
Abstract | Conversational agents such as Cortana, Alexa and Siri are continuously working on increasing their capabilities by adding new domains. The support of a new domain includes the design and development of a number of NLU components for domain classification, intents classification and slots tagging (including named entity recognition). Each component only performs well when trained on a large amount of labeled data. Second, these components are deployed on limited-memory devices which requires some model compression. Third, for some domains such as the health domain, it is hard to find a single training data set that covers all the required slot types. To overcome these mentioned problems, we present a multi-task transformer-based neural architecture for slot tagging. We consider the training of a slot tagger using multiple data sets covering different slot types as a multi-task learning problem. The experimental results on the biomedical domain have shown that the proposed approach outperforms the previous state-of-the-art systems for slot tagging on the different benchmark biomedical datasets in terms of (time and memory) efficiency and effectiveness. The output slot tagger can be used by the conversational agent to better identify entities in the input utterances. |
Tasks | Model Compression, Multi-Task Learning, Named Entity Recognition |
Published | 2020-01-24 |
URL | https://arxiv.org/abs/2001.08904v1 |
https://arxiv.org/pdf/2001.08904v1.pdf | |
PWC | https://paperswithcode.com/paper/mt-bioner-multi-task-learning-for-biomedical |
Repo | |
Framework | |
Learning When and Where to Zoom with Deep Reinforcement Learning
Title | Learning When and Where to Zoom with Deep Reinforcement Learning |
Authors | Burak Uzkent, Stefano Ermon |
Abstract | While high resolution images contain semantically more useful information than their lower resolution counterparts, processing them is computationally more expensive, and in some applications, e.g. remote sensing, they can be much more expensive to acquire. For these reasons, it is desirable to develop an automatic method to selectively use high resolution data when necessary while maintaining accuracy and reducing acquisition/run-time cost. In this direction, we propose PatchDrop a reinforcement learning approach to dynamically identify when and where to use/acquire high resolution data conditioned on the paired, cheap, low resolution images. We conduct experiments on CIFAR10, CIFAR100, ImageNet and fMoW datasets where we use significantly less high resolution data while maintaining similar accuracy to models which use full high resolution images. |
Tasks | |
Published | 2020-03-01 |
URL | https://arxiv.org/abs/2003.00425v1 |
https://arxiv.org/pdf/2003.00425v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-when-and-where-to-zoom-with-deep |
Repo | |
Framework | |
DeepFL-IQA: Weak Supervision for Deep IQA Feature Learning
Title | DeepFL-IQA: Weak Supervision for Deep IQA Feature Learning |
Authors | Hanhe Lin, Vlad Hosu, Dietmar Saupe |
Abstract | Multi-level deep-features have been driving state-of-the-art methods for aesthetics and image quality assessment (IQA). However, most IQA benchmarks are comprised of artificially distorted images, for which features derived from ImageNet under-perform. We propose a new IQA dataset and a weakly supervised feature learning approach to train features more suitable for IQA of artificially distorted images. The dataset, KADIS-700k, is far more extensive than similar works, consisting of 140,000 pristine images, 25 distortions types, totaling 700k distorted versions. Our weakly supervised feature learning is designed as a multi-task learning type training, using eleven existing full-reference IQA metrics as proxies for differential mean opinion scores. We also introduce a benchmark database, KADID-10k, of artificially degraded images, each subjectively annotated by 30 crowd workers. We make use of our derived image feature vectors for (no-reference) image quality assessment by training and testing a shallow regression network on this database and five other benchmark IQA databases. Our method, termed DeepFL-IQA, performs better than other feature-based no-reference IQA methods and also better than all tested full-reference IQA methods on KADID-10k. For the other five benchmark IQA databases, DeepFL-IQA matches the performance of the best existing end-to-end deep learning-based methods on average. |
Tasks | Image Quality Assessment, Multi-Task Learning, No-Reference Image Quality Assessment |
Published | 2020-01-20 |
URL | https://arxiv.org/abs/2001.08113v1 |
https://arxiv.org/pdf/2001.08113v1.pdf | |
PWC | https://paperswithcode.com/paper/deepfl-iqa-weak-supervision-for-deep-iqa |
Repo | |
Framework | |
On the Search for Feedback in Reinforcement Learning
Title | On the Search for Feedback in Reinforcement Learning |
Authors | Ran Wang, Karthikeya S. Parunandi, Dan Yu, Dileep Kalathil, Suman Chakravorty |
Abstract | This paper addresses the problem of learning the optimal feedback policy for a nonlinear stochastic dynamical system with continuous state space, continuous action space and unknown dynamics. Feedback policies are complex objects that typically need a large dimensional parametrization, which makes Reinforcement Learning algorithms that search for an optimum in this large parameter space, sample inefficient and subject to high variance. We propose a “decoupling” principle that drastically reduces the feedback parameter space while still remaining near-optimal to the fourth-order in a small noise parameter. Based on this principle, we propose a decoupled data-based control (D2C) algorithm that addresses the stochastic control problem: first, an open-loop deterministic trajectory optimization problem is solved using a black-box simulation model of the dynamical system. Then, a linear closed-loop control is developed around this nominal trajectory using only a simulation model. Empirical evidence suggests significant reduction in training time, as well as the training variance, compared to other state of the art Reinforcement Learning algorithms. |
Tasks | |
Published | 2020-02-21 |
URL | https://arxiv.org/abs/2002.09478v1 |
https://arxiv.org/pdf/2002.09478v1.pdf | |
PWC | https://paperswithcode.com/paper/on-the-search-for-feedback-in-reinforcement |
Repo | |
Framework | |
Gradient Surgery for Multi-Task Learning
Title | Gradient Surgery for Multi-Task Learning |
Authors | Tianhe Yu, Saurabh Kumar, Abhishek Gupta, Sergey Levine, Karol Hausman, Chelsea Finn |
Abstract | While deep learning and deep reinforcement learning (RL) systems have demonstrated impressive results in domains such as image classification, game playing, and robotic control, data efficiency remains a major challenge. Multi-task learning has emerged as a promising approach for sharing structure across multiple tasks to enable more efficient learning. However, the multi-task setting presents a number of optimization challenges, making it difficult to realize large efficiency gains compared to learning tasks independently. The reasons why multi-task learning is so challenging compared to single-task learning are not fully understood. In this work, we identify a set of three conditions of the multi-task optimization landscape that cause detrimental gradient interference, and develop a simple yet general approach for avoiding such interference between task gradients. We propose a form of gradient surgery that projects a task’s gradient onto the normal plane of the gradient of any other task that has a conflicting gradient. On a series of challenging multi-task supervised and multi-task RL problems, this approach leads to substantial gains in efficiency and performance. Further, it is model-agnostic and can be combined with previously-proposed multi-task architectures for enhanced performance. |
Tasks | Image Classification, Multi-Task Learning |
Published | 2020-01-19 |
URL | https://arxiv.org/abs/2001.06782v1 |
https://arxiv.org/pdf/2001.06782v1.pdf | |
PWC | https://paperswithcode.com/paper/gradient-surgery-for-multi-task-learning-1 |
Repo | |
Framework | |