April 2, 2020

3001 words 15 mins read

Paper Group ANR 166

Paper Group ANR 166

Beyond Near- and Long-Term: Towards a Clearer Account of Research Priorities in AI Ethics and Society. Extended Stochastic Gradient MCMC for Large-Scale Bayesian Variable Selection. Learning Near Optimal Policies with Low Inherent Bellman Error. Monte Carlo Anti-Differentiation for Approximate Weighted Model Integration. Bridge the Domain Gap Betwe …

Beyond Near- and Long-Term: Towards a Clearer Account of Research Priorities in AI Ethics and Society

Title Beyond Near- and Long-Term: Towards a Clearer Account of Research Priorities in AI Ethics and Society
Authors Carina Prunkl, Jess Whittlestone
Abstract One way of carving up the broad “AI ethics and society” research space that has emerged in recent years is to distinguish between “near-term” and “long-term” research. While such ways of breaking down the research space can be useful, we put forward several concerns about the near/long-term distinction gaining too much prominence in how research questions and priorities are framed. We highlight some ambiguities and inconsistencies in how the distinction is used, and argue that while there are differing priorities within this broad research community, these differences are not well-captured by the near/long-term distinction. We unpack the near/long-term distinction into four different dimensions, and propose some ways that researchers can communicate more clearly about their work and priorities using these dimensions. We suggest that moving towards a more nuanced conversation about research priorities can help establish new opportunities for collaboration, aid the development of more consistent and coherent research agendas, and enable identification of previously neglected research areas.
Tasks
Published 2020-01-13
URL https://arxiv.org/abs/2001.04335v2
PDF https://arxiv.org/pdf/2001.04335v2.pdf
PWC https://paperswithcode.com/paper/beyond-near-and-long-term-towards-a-clearer
Repo
Framework

Extended Stochastic Gradient MCMC for Large-Scale Bayesian Variable Selection

Title Extended Stochastic Gradient MCMC for Large-Scale Bayesian Variable Selection
Authors Qifan Song, Yan Sun, Mao Ye, Faming Liang
Abstract Stochastic gradient Markov chain Monte Carlo (MCMC) algorithms have received much attention in Bayesian computing for big data problems, but they are only applicable to a small class of problems for which the parameter space has a fixed dimension and the log-posterior density is differentiable with respect to the parameters. This paper proposes an extended stochastic gradient MCMC lgoriathm which, by introducing appropriate latent variables, can be applied to more general large-scale Bayesian computing problems, such as those involving dimension jumping and missing data. Numerical studies show that the proposed algorithm is highly scalable and much more efficient than traditional MCMC algorithms. The proposed algorithms have much alleviated the pain of Bayesian methods in big data computing.
Tasks
Published 2020-02-07
URL https://arxiv.org/abs/2002.02919v1
PDF https://arxiv.org/pdf/2002.02919v1.pdf
PWC https://paperswithcode.com/paper/extended-stochastic-gradient-mcmc-for-large
Repo
Framework

Learning Near Optimal Policies with Low Inherent Bellman Error

Title Learning Near Optimal Policies with Low Inherent Bellman Error
Authors Andrea Zanette, Alessandro Lazaric, Mykel Kochenderfer, Emma Brunskill
Abstract We study the exploration problem with approximate linear action-value functions in episodic reinforcement learning under the notion of low inherent Bellman error, a condition normally employed to show convergence of approximate value iteration. First we relate this condition to other common frameworks and show that it is strictly more general than the low rank (or linear) MDP assumption of prior work. Second we provide an algorithm with a high probability regret bound $\widetilde O(\sum_{t=1}^H d_t \sqrt{K} + \sum_{t=1}^H \sqrt{d_t} \IBE K)$ where $H$ is the horizon, $K$ is the number of episodes, $\IBE$ is the value if the inherent Bellman error and $d_t$ is the feature dimension at timestep $t$. In addition, we show that the result is unimprovable beyond constants and logs by showing a matching lower bound. This has two important consequences: 1) the algorithm has the optimal statistical rate for this setting which is more general than prior work on low-rank MDPs 2) the lack of closedness (measured by the inherent Bellman error) is only amplified by $\sqrt{d_t}$ despite working in the online setting. Finally, the algorithm reduces to the celebrated \textsc{LinUCB} when $H=1$ but with a different choice of the exploration parameter that allows handling misspecified contextual linear bandits. While computational tractability questions remain open for the MDP setting, this enriches the class of MDPs with a linear representation for the action-value function where statistically efficient reinforcement learning is possible.
Tasks
Published 2020-02-29
URL https://arxiv.org/abs/2003.00153v2
PDF https://arxiv.org/pdf/2003.00153v2.pdf
PWC https://paperswithcode.com/paper/learning-near-optimal-policies-with-low
Repo
Framework

Monte Carlo Anti-Differentiation for Approximate Weighted Model Integration

Title Monte Carlo Anti-Differentiation for Approximate Weighted Model Integration
Authors Pedro Zuidberg Dos Martires, Samuel Kolb
Abstract Probabilistic inference in the hybrid domain, i.e. inference over discrete-continuous domains, requires tackling two well known #P-hard problems 1)~weighted model counting (WMC) over discrete variables and 2)~integration over continuous variables. For both of these problems inference techniques have been developed separately in order to manage their #P-hardness, such as knowledge compilation for WMC and Monte Carlo (MC) methods for (approximate) integration in the continuous domain. Weighted model integration (WMI), the extension of WMC to the hybrid domain, has been proposed as a formalism to study probabilistic inference over discrete and continuous variables alike. Recently developed WMI solvers have focused on exploiting structure in WMI problems, for which they rely on symbolic integration to find the primitive of an integrand, i.e. to perform anti-differentiation. To combine these advances with state-of-the-art Monte Carlo integration techniques, we introduce \textit{Monte Carlo anti-differentiation} (MCAD), which computes MC approximations of anti-derivatives. In our empirical evaluation we substitute the exact symbolic integration backend in an existing WMI solver with an MCAD backend. Our experiments show that that equipping existing WMI solvers with MCAD yields a fast yet reliable approximate inference scheme.
Tasks
Published 2020-01-13
URL https://arxiv.org/abs/2001.04566v1
PDF https://arxiv.org/pdf/2001.04566v1.pdf
PWC https://paperswithcode.com/paper/monte-carlo-anti-differentiation-for
Repo
Framework

Bridge the Domain Gap Between Ultra-wide-field and Traditional Fundus Images via Adversarial Domain Adaptation

Title Bridge the Domain Gap Between Ultra-wide-field and Traditional Fundus Images via Adversarial Domain Adaptation
Authors Lie Ju, Xin Wang, Quan Zhou, Hu Zhu, Mehrtash Harandi, Paul Bonnington, Tom Drummond, Zongyuan Ge
Abstract For decades, advances in retinal imaging technology have enabled effective diagnosis and management of retinal disease using fundus cameras. Recently, ultra-wide-field (UWF) fundus imaging by Optos camera is gradually put into use because of its broader insights on fundus for some lesions that are not typically seen in traditional fundus images. Research on traditional fundus images is an active topic but studies on UWF fundus images are few. One of the most important reasons is that UWF fundus images are hard to obtain. In this paper, for the first time, we explore domain adaptation from the traditional fundus to UWF fundus images. We propose a flexible framework to bridge the domain gap between two domains and co-train a UWF fundus diagnosis model by pseudo-labelling and adversarial learning. We design a regularisation technique to regulate the domain adaptation. Also, we apply MixUp to overcome the over-fitting issue from incorrect generated pseudo-labels. Our experimental results on either single or both domains demonstrate that the proposed method can well adapt and transfer the knowledge from traditional fundus images to UWF fundus images and improve the performance of retinal disease recognition.
Tasks Domain Adaptation
Published 2020-03-23
URL https://arxiv.org/abs/2003.10042v2
PDF https://arxiv.org/pdf/2003.10042v2.pdf
PWC https://paperswithcode.com/paper/bridge-the-domain-gap-between-ultra-wide
Repo
Framework

Sharp Asymptotics and Optimal Performance for Inference in Binary Models

Title Sharp Asymptotics and Optimal Performance for Inference in Binary Models
Authors Hossein Taheri, Ramtin Pedarsani, Christos Thrampoulidis
Abstract We study convex empirical risk minimization for high-dimensional inference in binary models. Our first result sharply predicts the statistical performance of such estimators in the linear asymptotic regime under isotropic Gaussian features. Importantly, the predictions hold for a wide class of convex loss functions, which we exploit in order to prove a bound on the best achievable performance among them. Notably, we show that the proposed bound is tight for popular binary models (such as Signed, Logistic or Probit), by constructing appropriate loss functions that achieve it. More interestingly, for binary linear classification under the Logistic and Probit models, we prove that the performance of least-squares is no worse than 0.997 and 0.98 times the optimal one. Numerical simulations corroborate our theoretical findings and suggest they are accurate even for relatively small problem dimensions.
Tasks
Published 2020-02-17
URL https://arxiv.org/abs/2002.07284v2
PDF https://arxiv.org/pdf/2002.07284v2.pdf
PWC https://paperswithcode.com/paper/sharp-asymptotics-and-optimal-performance-for
Repo
Framework

Bayesian Learning of Causal Relationships for System Reliability

Title Bayesian Learning of Causal Relationships for System Reliability
Authors Xuewen Yu, Jim Q. Smith, Linda Nichols
Abstract Causal theory is now widely developed with many applications to medicine and public health. However within the discipline of reliability, although causation is a key concept in this field, there has been much less theoretical attention. In this paper, we will demonstrate how some aspects of established causal methodology can be translated via trees, and more specifically chain event graphs, into domain of reliability theory to help the probability modeling of failures. We further show how various domain specific concepts of causality particular to reliability can be imported into more generic causal algebras and so demonstrate how these disciplines can inform each other. This paper is informed by a detailed analysis of maintenance records associated with a large electrical distribution company. Causal hypotheses embedded within these natural language texts are extracted and analyzed using the new graphical framework we introduced here.
Tasks
Published 2020-02-14
URL https://arxiv.org/abs/2002.06084v1
PDF https://arxiv.org/pdf/2002.06084v1.pdf
PWC https://paperswithcode.com/paper/bayesian-learning-of-causal-relationships-for
Repo
Framework

Wasserstein-based Graph Alignment

Title Wasserstein-based Graph Alignment
Authors Hermina Petric Maretic, Mireille El Gheche, Matthias Minder, Giovanni Chierchia, Pascal Frossard
Abstract We propose a novel method for comparing non-aligned graphs of different sizes, based on the Wasserstein distance between graph signal distributions induced by the respective graph Laplacian matrices. Specifically, we cast a new formulation for the one-to-many graph alignment problem, which aims at matching a node in the smaller graph with one or more nodes in the larger graph. By integrating optimal transport in our graph comparison framework, we generate both a structurally-meaningful graph distance, and a signal transportation plan that models the structure of graph data. The resulting alignment problem is solved with stochastic gradient descent, where we use a novel Dykstra operator to ensure that the solution is a one-to-many (soft) assignment matrix. We demonstrate the performance of our novel framework on graph alignment and graph classification, and we show that our method leads to significant improvements with respect to the state-of-the-art algorithms for each of these tasks.
Tasks Graph Classification
Published 2020-03-12
URL https://arxiv.org/abs/2003.06048v1
PDF https://arxiv.org/pdf/2003.06048v1.pdf
PWC https://paperswithcode.com/paper/wasserstein-based-graph-alignment
Repo
Framework

Distance Metric Learning for Graph Structured Data

Title Distance Metric Learning for Graph Structured Data
Authors Tomoki Yoshida, Ichiro Takeuchi, Masayuki Karasuyama
Abstract Graphs are versatile tools for representing structured data. Therefore, a variety of machine learning methods have been studied for graph data analysis. Although many of those learning methods depend on the measurement of differences between input graphs, defining an appropriate distance metric for a graph remains a controversial issue. Hence, we propose a supervised distance metric learning method for the graph classification problem. Our method, named interpretable graph metric learning (IGML), learns discriminative metrics in a subgraph-based feature space, which has a strong graph representation capability. By introducing a sparsity-inducing penalty on a weight of each subgraph, IGML can identify a small number of important subgraphs that can provide insight about the given classification task. Since our formulation has a large number of optimization variables, an efficient algorithm is also proposed by using pruning techniques based on safe screening and working set selection methods. An important property of IGML is that the optimality of the solution is guaranteed because the problem is formulated as a convex problem and our pruning strategies only discard unnecessary subgraphs. Further, we show that IGML is also applicable to other structured data such as item-set and sequence data, and that it can incorporate vertex-label similarity by using a transportation-based subgraph feature. We empirically evaluate the computational efficiency and classification performance on several benchmark datasets and show some illustrative examples demonstrating that IGML identifies important subgraphs from a given graph dataset.
Tasks Graph Classification, Metric Learning
Published 2020-02-03
URL https://arxiv.org/abs/2002.00727v1
PDF https://arxiv.org/pdf/2002.00727v1.pdf
PWC https://paperswithcode.com/paper/distance-metric-learning-for-graph-structured
Repo
Framework

Clustering with Fast, Automated and Reproducible assessment applied to longitudinal neural tracking

Title Clustering with Fast, Automated and Reproducible assessment applied to longitudinal neural tracking
Authors Hanlin Zhu, Xue Li, Liuyang Sun, Fei He, Zhengtuo Zhao, Lan Luan, Ngoc Mai Tran, Chong Xie
Abstract Across many areas, from neural tracking to database entity resolution, manual assessment of clusters by human experts presents a bottleneck in rapid development of scalable and specialized clustering methods. To solve this problem we develop C-FAR, a novel method for Fast, Automated and Reproducible assessment of multiple hierarchical clustering algorithms simultaneously. Our algorithm takes any number of hierarchical clustering trees as input, then strategically queries pairs for human feedback, and outputs an optimal clustering among those nominated by these trees. While it is applicable to large dataset in any domain that utilizes pairwise comparisons for assessment, our flagship application is the cluster aggregation step in spike-sorting, the task of assigning waveforms (spikes) in recordings to neurons. On simulated data of 96 neurons under adverse conditions, including drifting and 25% blackout, our algorithm produces near-perfect tracking relative to the ground truth. Our runtime scales linearly in the number of input trees, making it a competitive computational tool. These results indicate that C-FAR is highly suitable as a model selection and assessment tool in clustering tasks.
Tasks Entity Resolution, Model Selection
Published 2020-03-19
URL https://arxiv.org/abs/2003.08533v1
PDF https://arxiv.org/pdf/2003.08533v1.pdf
PWC https://paperswithcode.com/paper/clustering-with-fast-automated-and
Repo
Framework

Class-Specific Blind Deconvolutional Phase Retrieval Under a Generative Prior

Title Class-Specific Blind Deconvolutional Phase Retrieval Under a Generative Prior
Authors Fahad Shamshad, Ali Ahmed
Abstract In this paper, we consider the highly ill-posed problem of jointly recovering two real-valued signals from the phaseless measurements of their circular convolution. The problem arises in various imaging modalities such as Fourier ptychography, X-ray crystallography, and in visible light communication. We propose to solve this inverse problem using alternating gradient descent algorithm under two pretrained deep generative networks as priors; one is trained on sharp images and the other on blur kernels. The proposed recovery algorithm strives to find a sharp image and a blur kernel in the range of the respective pre-generators that \textit{best} explain the forward measurement model. In doing so, we are able to reconstruct quality image estimates. Moreover, the numerics show that the proposed approach performs well on the challenging measurement models that reflect the physically realizable imaging systems and is also robust to noise
Tasks
Published 2020-02-28
URL https://arxiv.org/abs/2002.12578v1
PDF https://arxiv.org/pdf/2002.12578v1.pdf
PWC https://paperswithcode.com/paper/class-specific-blind-deconvolutional-phase
Repo
Framework

Learning Implicit Surface Light Fields

Title Learning Implicit Surface Light Fields
Authors Michael Oechsle, Michael Niemeyer, Lars Mescheder, Thilo Strauss, Andreas Geiger
Abstract Implicit representations of 3D objects have recently achieved impressive results on learning-based 3D reconstruction tasks. While existing works use simple texture models to represent object appearance, photo-realistic image synthesis requires reasoning about the complex interplay of light, geometry and surface properties. In this work, we propose a novel implicit representation for capturing the visual appearance of an object in terms of its surface light field. In contrast to existing representations, our implicit model represents surface light fields in a continuous fashion and independent of the geometry. Moreover, we condition the surface light field with respect to the location and color of a small light source. Compared to traditional surface light field models, this allows us to manipulate the light source and relight the object using environment maps. We further demonstrate the capabilities of our model to predict the visual appearance of an unseen object from a single real RGB image and corresponding 3D shape information. As evidenced by our experiments, our model is able to infer rich visual appearance including shadows and specular reflections. Finally, we show that the proposed representation can be embedded into a variational auto-encoder for generating novel appearances that conform to the specified illumination conditions.
Tasks 3D Reconstruction, Image Generation
Published 2020-03-27
URL https://arxiv.org/abs/2003.12406v1
PDF https://arxiv.org/pdf/2003.12406v1.pdf
PWC https://paperswithcode.com/paper/learning-implicit-surface-light-fields
Repo
Framework

StrokeCoder: Path-Based Image Generation from Single Examples using Transformers

Title StrokeCoder: Path-Based Image Generation from Single Examples using Transformers
Authors Sabine Wieluch, Friedhelm Schwenker
Abstract This paper demonstrates how a Transformer Neural Network can be used to learn a Generative Model from a single path-based example image. We further show how a data set can be generated from the example image and how the model can be used to generate a large set of deviated images, which still represent the original image’s style and concept.
Tasks Image Generation
Published 2020-03-26
URL https://arxiv.org/abs/2003.11958v1
PDF https://arxiv.org/pdf/2003.11958v1.pdf
PWC https://paperswithcode.com/paper/strokecoder-path-based-image-generation-from
Repo
Framework

Cycle Text-To-Image GAN with BERT

Title Cycle Text-To-Image GAN with BERT
Authors Trevor Tsue, Samir Sen, Jason Li
Abstract We explore novel approaches to the task of image generation from their respective captions, building on state-of-the-art GAN architectures. Particularly, we baseline our models with the Attention-based GANs that learn attention mappings from words to image features. To better capture the features of the descriptions, we then built a novel cyclic design that learns an inverse function to maps the image back to original caption. Additionally, we incorporated recently developed BERT pretrained word embeddings as our initial text featurizer and observe a noticeable improvement in qualitative and quantitative performance compared to the Attention GAN baseline.
Tasks Image Generation, Word Embeddings
Published 2020-03-26
URL https://arxiv.org/abs/2003.12137v1
PDF https://arxiv.org/pdf/2003.12137v1.pdf
PWC https://paperswithcode.com/paper/cycle-text-to-image-gan-with-bert
Repo
Framework

Boolean learning under noise-perturbations in hardware neural networks

Title Boolean learning under noise-perturbations in hardware neural networks
Authors Louis Andreoli, Xavier Porte, Stéphane Chrétien, Maxime Jacquot, Laurent Larger, Daniel Brunner
Abstract A high efficiency hardware integration of neural networks benefits from realizing nonlinearity, network connectivity and learning fully in a physical substrate. Multiple systems have recently implemented some or all of these operations, yet the focus was placed on addressing technological challenges. Fundamental questions regarding learning in hardware neural networks remain largely unexplored. Noise in particular is unavoidable in such architectures, and here we investigate its interaction with a learning algorithm using an opto-electronic recurrent neural network. We find that noise strongly modifies the system’s path during convergence, and surprisingly fully decorrelates the final readout weight matrices. This highlights the importance of understanding architecture, noise and learning algorithm as interacting players, and therefore identifies the need for mathematical tools for noisy, analogue system optimization.
Tasks
Published 2020-03-27
URL https://arxiv.org/abs/2003.12319v1
PDF https://arxiv.org/pdf/2003.12319v1.pdf
PWC https://paperswithcode.com/paper/boolean-learning-under-noise-perturbations-in
Repo
Framework
comments powered by Disqus