January 31, 2020

2897 words 14 mins read

Paper Group AWR 392

Paper Group AWR 392

Traffic map prediction using UNet based deep convolutional neural network. OmniNet: A unified architecture for multi-modal multi-task learning. Composable Effects for Flexible and Accelerated Probabilistic Programming in NumPyro. nnU-Net: Breaking the Spell on Successful Medical Image Segmentation. Real numbers, data science and chaos: How to fit a …

Traffic map prediction using UNet based deep convolutional neural network

Title Traffic map prediction using UNet based deep convolutional neural network
Authors Sungbin Choi
Abstract This paper describes our UNet based deep convolutional neural network approach on the Traffic4cast challenge 2019. Challenges task is to predict future traffic flow volume, heading and speed on high resolution whole city map. We used UNet based deep convolutional neural network to train predictive model for the short term traffic forecast. On each convolution block, layers are densely connected with subsequent layers like a DenseNet. Trained and evaluated on the real world data set collected from three distinct cities in the world, our method achieved best performance in this challenge.
Tasks
Published 2019-11-25
URL https://arxiv.org/abs/1912.05288v1
PDF https://arxiv.org/pdf/1912.05288v1.pdf
PWC https://paperswithcode.com/paper/traffic-map-prediction-using-unet-based-deep
Repo https://github.com/SungbinChoi/traffic4cast2019
Framework tf

OmniNet: A unified architecture for multi-modal multi-task learning

Title OmniNet: A unified architecture for multi-modal multi-task learning
Authors Subhojeet Pramanik, Priyanka Agrawal, Aman Hussain
Abstract Transformer is a popularly used neural network architecture, especially for language understanding. We introduce an extended and unified architecture which can be used for tasks involving a variety of modalities like image, text, videos, etc. We propose a spatio-temporal cache mechanism that enables learning spatial dimension of the input in addition to the hidden states corresponding to the temporal input sequence. The proposed architecture further enables a single model to support tasks with multiple input modalities as well as asynchronous multi-task learning, thus we refer to it as OmniNet. For example, a single instance of OmniNet can concurrently learn to perform the tasks of part-of-speech tagging, image captioning, visual question answering and video activity recognition. We demonstrate that training these four tasks together results in about three times compressed model while retaining the performance in comparison to training them individually. We also show that using this neural network pre-trained on some modalities assists in learning an unseen task. This illustrates the generalization capacity of the self-attention mechanism on the spatio-temporal cache present in OmniNet.
Tasks Image Captioning, Multi-Task Learning, Part-Of-Speech Tagging, Question Answering, Visual Question Answering
Published 2019-07-17
URL https://arxiv.org/abs/1907.07804v1
PDF https://arxiv.org/pdf/1907.07804v1.pdf
PWC https://paperswithcode.com/paper/omninet-a-unified-architecture-for-multi
Repo https://github.com/subho406/OmniNet
Framework pytorch

Composable Effects for Flexible and Accelerated Probabilistic Programming in NumPyro

Title Composable Effects for Flexible and Accelerated Probabilistic Programming in NumPyro
Authors Du Phan, Neeraj Pradhan, Martin Jankowiak
Abstract NumPyro is a lightweight library that provides an alternate NumPy backend to the Pyro probabilistic programming language with the same modeling interface, language primitives and effect handling abstractions. Effect handlers allow Pyro’s modeling API to be extended to NumPyro despite its being built atop a fundamentally different JAX-based functional backend. In this work, we demonstrate the power of composing Pyro’s effect handlers with the program transformations that enable hardware acceleration, automatic differentiation, and vectorization in JAX. In particular, NumPyro provides an iterative formulation of the No-U-Turn Sampler (NUTS) that can be end-to-end JIT compiled, yielding an implementation that is much faster than existing alternatives in both the small and large dataset regimes.
Tasks Probabilistic Programming
Published 2019-12-24
URL https://arxiv.org/abs/1912.11554v1
PDF https://arxiv.org/pdf/1912.11554v1.pdf
PWC https://paperswithcode.com/paper/composable-effects-for-flexible-and
Repo https://github.com/pyro-ppl/numpyro
Framework jax

nnU-Net: Breaking the Spell on Successful Medical Image Segmentation

Title nnU-Net: Breaking the Spell on Successful Medical Image Segmentation
Authors Fabian Isensee, Jens Petersen, Simon A. A. Kohl, Paul F. Jäger, Klaus H. Maier-Hein
Abstract Fueled by the diversity of datasets, semantic segmentation is a popular subfield in medical image analysis with a vast number of new methods being proposed each year. This ever-growing jungle of methodologies, however, becomes increasingly impenetrable. At the same time, many proposed methods fail to generalize beyond the experiments they were demonstrated on, thus hampering the process of developing a segmentation algorithm on a new dataset. Here we present nnU-Net (‘no-new-Net’), a framework that automatically adapts itself to any given new dataset. While this process was completely human-driven so far, we make a first attempt to automate necessary adaptations such as preprocessing, the exact patch size, batch size, and inference settings based on the properties of a given dataset. Remarkably, nnU-Net strips away the architectural bells and whistles that are typically proposed in the literature and relies on just a simple U-Net architecture embedded in a robust training scheme. Out of the box, nnU-Net achieves state of the art performance on six well-established segmentation challenges. Source code is available at https://github.com/MIC-DKFZ/nnunet.
Tasks Medical Image Segmentation, Semantic Segmentation
Published 2019-04-17
URL http://arxiv.org/abs/1904.08128v1
PDF http://arxiv.org/pdf/1904.08128v1.pdf
PWC https://paperswithcode.com/paper/nnu-net-breaking-the-spell-on-successful
Repo https://github.com/NeuroAI-HD/HD-GLIO-AUTO
Framework none

Real numbers, data science and chaos: How to fit any dataset with a single parameter

Title Real numbers, data science and chaos: How to fit any dataset with a single parameter
Authors Laurent Boué
Abstract We show how any dataset of any modality (time-series, images, sound…) can be approximated by a well-behaved (continuous, differentiable…) scalar function with a single real-valued parameter. Building upon elementary concepts from chaos theory, we adopt a pedagogical approach demonstrating how to adjust this parameter in order to achieve arbitrary precision fit to all samples of the data. Targeting an audience of data scientists with a taste for the curious and unusual, the results presented here expand on previous similar observations regarding expressiveness power and generalization of machine learning models.
Tasks Time Series
Published 2019-04-28
URL http://arxiv.org/abs/1904.12320v1
PDF http://arxiv.org/pdf/1904.12320v1.pdf
PWC https://paperswithcode.com/paper/real-numbers-data-science-and-chaos-how-to
Repo https://github.com/Ranlot/single-parameter-fit
Framework none

Bias Correction of Learned Generative Models using Likelihood-Free Importance Weighting

Title Bias Correction of Learned Generative Models using Likelihood-Free Importance Weighting
Authors Aditya Grover, Jiaming Song, Alekh Agarwal, Kenneth Tran, Ashish Kapoor, Eric Horvitz, Stefano Ermon
Abstract A learned generative model often produces biased statistics relative to the underlying data distribution. A standard technique to correct this bias is importance sampling, where samples from the model are weighted by the likelihood ratio under model and true distributions. When the likelihood ratio is unknown, it can be estimated by training a probabilistic classifier to distinguish samples from the two distributions. We employ this likelihood-free importance weighting method to correct for the bias in generative models. We find that this technique consistently improves standard goodness-of-fit metrics for evaluating the sample quality of state-of-the-art deep generative models, suggesting reduced bias. Finally, we demonstrate its utility on representative applications in a) data augmentation for classification using generative adversarial networks, and b) model-based policy evaluation using off-policy data.
Tasks Data Augmentation
Published 2019-06-23
URL https://arxiv.org/abs/1906.09531v2
PDF https://arxiv.org/pdf/1906.09531v2.pdf
PWC https://paperswithcode.com/paper/bias-correction-of-learned-generative-models
Repo https://github.com/aditya-grover/bias-correction-generative
Framework pytorch

EVDodgeNet: Deep Dynamic Obstacle Dodging with Event Cameras

Title EVDodgeNet: Deep Dynamic Obstacle Dodging with Event Cameras
Authors Nitin J. Sanket, Chethan M. Parameshwara, Chahat Deep Singh, Ashwin V. Kuruttukulam, Cornelia Fermüller, Davide Scaramuzza, Yiannis Aloimonos
Abstract Dynamic obstacle avoidance on quadrotors requires low latency. A class of sensors that are particularly suitable for such scenarios are event cameras. In this paper, we present a deep learning – based solution for dodging multiple dynamic obstacles on a quadrotor with a single event camera and on-board computation. Our approach uses a series of shallow neural networks for estimating both the ego-motion and the motion of independently moving objects. The networks are trained in simulation and directly transfer to the real world without any fine-tuning or retraining. We successfully evaluate and demonstrate the proposed approach in many real-world experiments with obstacles of different shapes and sizes, achieving an overall success rate of 70% including objects of unknown shape and a low light testing scenario. To our knowledge, this is the first deep learning – based solution to the problem of dynamic obstacle avoidance using event cameras on a quadrotor. Finally, we also extend our work to the pursuit task by merely reversing the control policy, proving that our navigation stack can cater to different scenarios.
Tasks Motion Estimation
Published 2019-06-07
URL https://arxiv.org/abs/1906.02919v3
PDF https://arxiv.org/pdf/1906.02919v3.pdf
PWC https://paperswithcode.com/paper/evdodge-embodied-ai-for-high-speed-dodging-on
Repo https://github.com/prgumd/EVDodge
Framework none

Provable Certificates for Adversarial Examples: Fitting a Ball in the Union of Polytopes

Title Provable Certificates for Adversarial Examples: Fitting a Ball in the Union of Polytopes
Authors Matt Jordan, Justin Lewis, Alexandros G. Dimakis
Abstract We propose a novel method for computing exact pointwise robustness of deep neural networks for all convex $\ell_p$ norms. Our algorithm, GeoCert, finds the largest $\ell_p$ ball centered at an input point $x_0$, within which the output class of a given neural network with ReLU nonlinearities remains unchanged. We relate the problem of computing pointwise robustness of these networks to that of computing the maximum norm ball with a fixed center that can be contained in a non-convex polytope. This is a challenging problem in general, however we show that there exists an efficient algorithm to compute this for polyhedral complices. Further we show that piecewise linear neural networks partition the input space into a polyhedral complex. Our algorithm has the ability to almost immediately output a nontrivial lower bound to the pointwise robustness which is iteratively improved until it ultimately becomes tight. We empirically show that our approach generates distance lower bounds that are tighter compared to prior work, under moderate time constraints.
Tasks
Published 2019-03-20
URL https://arxiv.org/abs/1903.08778v2
PDF https://arxiv.org/pdf/1903.08778v2.pdf
PWC https://paperswithcode.com/paper/provable-certificates-for-adversarial
Repo https://github.com/revbucket/geometric-certificates
Framework pytorch

Bridging the domain gap in cross-lingual document classification

Title Bridging the domain gap in cross-lingual document classification
Authors Guokun Lai, Barlas Oguz, Yiming Yang, Veselin Stoyanov
Abstract The scarcity of labeled training data often prohibits the internationalization of NLP models to multiple languages. Recent developments in cross-lingual understanding (XLU) has made progress in this area, trying to bridge the language barrier using language universal representations. However, even if the language problem was resolved, models trained in one language would not transfer to another language perfectly due to the natural domain drift across languages and cultures. We consider the setting of semi-supervised cross-lingual understanding, where labeled data is available in a source language (English), but only unlabeled data is available in the target language. We combine state-of-the-art cross-lingual methods with recently proposed methods for weakly supervised learning such as unsupervised pre-training and unsupervised data augmentation to simultaneously close both the language gap and the domain gap in XLU. We show that addressing the domain gap is crucial. We improve over strong baselines and achieve a new state-of-the-art for cross-lingual document classification.
Tasks Cross-Lingual Document Classification, Data Augmentation, Document Classification
Published 2019-09-16
URL https://arxiv.org/abs/1909.07009v2
PDF https://arxiv.org/pdf/1909.07009v2.pdf
PWC https://paperswithcode.com/paper/bridging-the-domain-gap-in-cross-lingual
Repo https://github.com/laiguokun/xlu-data
Framework none

Creative GANs for generating poems, lyrics, and metaphors

Title Creative GANs for generating poems, lyrics, and metaphors
Authors Asir Saeed, Suzana Ilić, Eva Zangerle
Abstract Generative models for text have substantially contributed to tasks like machine translation and language modeling, using maximum likelihood optimization (MLE). However, for creative text generation, where multiple outputs are possible and originality and uniqueness are encouraged, MLE falls short. Methods optimized for MLE lead to outputs that can be generic, repetitive and incoherent. In this work, we use a Generative Adversarial Network framework to alleviate this problem. We evaluate our framework on poetry, lyrics and metaphor datasets, each with widely different characteristics, and report better performance of our objective function over other generative models.
Tasks Language Modelling, Machine Translation, Text Generation
Published 2019-09-20
URL https://arxiv.org/abs/1909.09534v1
PDF https://arxiv.org/pdf/1909.09534v1.pdf
PWC https://paperswithcode.com/paper/creative-gans-for-generating-poems-lyrics-and
Repo https://github.com/Machine-Learning-Tokyo/Poetry-GAN
Framework none

Symbolic Graph Embedding using Frequent Pattern Mining

Title Symbolic Graph Embedding using Frequent Pattern Mining
Authors Blaz Škrlj, Jan Kralj, Nada Lavrač
Abstract Relational data mining is becoming ubiquitous in many fields of study. It offers insights into behaviour of complex, real-world systems which cannot be modeled directly using propositional learning. We propose Symbolic Graph Embedding (SGE), an algorithm aimed to learn symbolic node representations. Built on the ideas from the field of inductive logic programming, SGE first samples a given node’s neighborhood and interprets it as a transaction database, which is used for frequent pattern mining to identify logical conjuncts of items that co-occur frequently in a given context. Such patterns are in this work used as features to represent individual nodes, yielding interpretable, symbolic node embeddings. The proposed SGE approach on a venue classification task outperforms shallow node embedding methods such as DeepWalk, and performs similarly to metapath2vec, a black-box representation learner that can exploit node and edge types in a given graph. The proposed SGE approach performs especially well when small amounts of data are used for learning, scales to graphs with millions of nodes and edges, and can be run on an of-the-shelf laptop.
Tasks Graph Embedding
Published 2019-10-29
URL https://arxiv.org/abs/1910.13314v1
PDF https://arxiv.org/pdf/1910.13314v1.pdf
PWC https://paperswithcode.com/paper/191013314
Repo https://github.com/SkBlaz/SGE
Framework none

Fair Resource Allocation in Federated Learning

Title Fair Resource Allocation in Federated Learning
Authors Tian Li, Maziar Sanjabi, Ahmad Beirami, Virginia Smith
Abstract Federated learning involves training statistical models in massive, heterogeneous networks. Naively minimizing an aggregate loss function in such a network may disproportionately advantage or disadvantage some of the devices. In this work, we propose q-Fair Federated Learning (q-FFL), a novel optimization objective inspired by fair resource allocation in wireless networks that encourages a more fair (specifically, a more uniform) accuracy distribution across devices in federated networks. To solve q-FFL, we devise a communication-efficient method, q-FedAvg, that is suited to federated networks. We validate both the effectiveness of q-FFL and the efficiency of q-FedAvg on a suite of federated datasets with both convex and non-convex models, and show that q-FFL (along with q-FedAvg) outperforms existing baselines in terms of the resulting fairness, flexibility, and efficiency.
Tasks
Published 2019-05-25
URL https://arxiv.org/abs/1905.10497v2
PDF https://arxiv.org/pdf/1905.10497v2.pdf
PWC https://paperswithcode.com/paper/190510497
Repo https://github.com/litian96/fair_flearn
Framework none

Nonlinear generalization of the monotone single index model

Title Nonlinear generalization of the monotone single index model
Authors Zeljko Kereta, Timo Klock, Valeriya Naumova
Abstract Single index model is a powerful yet simple model, widely used in statistics, machine learning, and other scientific fields. It models the regression function as $g(<a,x>)$, where a is an unknown index vector and x are the features. This paper deals with a nonlinear generalization of this framework to allow for a regressor that uses multiple index vectors, adapting to local changes in the responses. To do so we exploit the conditional distribution over function-driven partitions, and use linear regression to locally estimate index vectors. We then regress by applying a kNN type estimator that uses a localized proxy of the geodesic metric. We present theoretical guarantees for estimation of local index vectors and out-of-sample prediction, and demonstrate the performance of our method with experiments on synthetic and real-world data sets, comparing it with state-of-the-art methods.
Tasks
Published 2019-02-24
URL https://arxiv.org/abs/1902.09024v2
PDF https://arxiv.org/pdf/1902.09024v2.pdf
PWC https://paperswithcode.com/paper/nonlinear-generalization-of-the-single-index
Repo https://github.com/soply/local_sim_experiments
Framework none

s-LWSR: Super Lightweight Super-Resolution Network

Title s-LWSR: Super Lightweight Super-Resolution Network
Authors Biao Li, Jiabin Liu, Bo Wang, Zhiquan Qi, Yong Shi
Abstract Deep learning (DL) architectures for superresolution (SR) normally contain tremendous parameters, which has been regarded as the crucial advantage for obtaining satisfying performance. However, with the widespread use of mobile phones for taking and retouching photos, this character greatly hampers the deployment of DL-SR models on the mobile devices. To address this problem, in this paper, we propose a super lightweight SR network: s-LWSR. There are mainly three contributions in our work. Firstly, in order to efficiently abstract features from the low resolution image, we build an information pool to mix multi-level information from the first half part of the pipeline. Accordingly, the information pool feeds the second half part with the combination of hierarchical features from the previous layers. Secondly, we employ a compression module to further decrease the size of parameters. Intensive analysis confirms its capacity of trade-off between model complexity and accuracy. Thirdly, by revealing the specific role of activation in deep models, we remove several activation layers in our SR model to retain more information for performance improvement. Extensive experiments show that our s-LWSR, with limited parameters and operations, can achieve similar performance to other cumbersome DL-SR methods.
Tasks Super-Resolution
Published 2019-09-24
URL https://arxiv.org/abs/1909.10774v1
PDF https://arxiv.org/pdf/1909.10774v1.pdf
PWC https://paperswithcode.com/paper/s-lwsr-super-lightweight-super-resolution
Repo https://github.com/Sudo-Biao/s-LWSR
Framework pytorch

A Syntax-aware Multi-task Learning Framework for Chinese Semantic Role Labeling

Title A Syntax-aware Multi-task Learning Framework for Chinese Semantic Role Labeling
Authors Qingrong Xia, Zhenghua Li, Min Zhang
Abstract Semantic role labeling (SRL) aims to identify the predicate-argument structure of a sentence. Inspired by the strong correlation between syntax and semantics, previous works pay much attention to improve SRL performance on exploiting syntactic knowledge, achieving significant results. Pipeline methods based on automatic syntactic trees and multi-task learning (MTL) approaches using standard syntactic trees are two common research orientations. In this paper, we adopt a simple unified span-based model for both span-based and word-based Chinese SRL as a strong baseline. Besides, we present a MTL framework that includes the basic SRL module and a dependency parser module. Different from the commonly used hard parameter sharing strategy in MTL, the main idea is to extract implicit syntactic representations from the dependency parser as external inputs for the basic SRL model. Experiments on the benchmarks of Chinese Proposition Bank 1.0 and CoNLL-2009 Chinese datasets show that our proposed framework can effectively improve the performance over the strong baselines. With the external BERT representations, our framework achieves new state-of-the-art 87.54 and 88.5 F1 scores on the two test data of the two benchmarks, respectively. In-depth analysis are conducted to gain more insights on the proposed framework and the effectiveness of syntax.
Tasks Multi-Task Learning, Semantic Role Labeling
Published 2019-11-12
URL https://arxiv.org/abs/1911.04641v1
PDF https://arxiv.org/pdf/1911.04641v1.pdf
PWC https://paperswithcode.com/paper/a-syntax-aware-multi-task-learning-framework-1
Repo https://github.com/KiroSummer/A_Syntax-aware_MTL_Framework_for_Chinese_SRL
Framework pytorch
comments powered by Disqus