October 19, 2019

3093 words 15 mins read

Paper Group ANR 219

Paper Group ANR 219

A Description Logic Framework for Commonsense Conceptual Combination Integrating Typicality, Probabilities and Cognitive Heuristics. A Decomposition-Based Many-Objective Evolutionary Algorithm with Local Iterative Update. Analysis of adversarial attacks against CNN-based image forgery detectors. Sea surface temperature prediction and reconstruction …

A Description Logic Framework for Commonsense Conceptual Combination Integrating Typicality, Probabilities and Cognitive Heuristics

Title A Description Logic Framework for Commonsense Conceptual Combination Integrating Typicality, Probabilities and Cognitive Heuristics
Authors Antonio Lieto, Gian Luca Pozzato
Abstract We propose a nonmonotonic Description Logic of typicality able to account for the phenomenon of concept combination of prototypical concepts. The proposed logic relies on the logic of typicality ALC TR, whose semantics is based on the notion of rational closure, as well as on the distributed semantics of probabilistic Description Logics, and is equipped with a cognitive heuristic used by humans for concept composition. We first extend the logic of typicality ALC TR by typicality inclusions whose intuitive meaning is that “there is probability p about the fact that typical Cs are Ds”. As in the distributed semantics, we define different scenarios containing only some typicality inclusions, each one having a suitable probability. We then focus on those scenarios whose probabilities belong to a given and fixed range, and we exploit such scenarios in order to ascribe typical properties to a concept C obtained as the combination of two prototypical concepts. We also show that reasoning in the proposed Description Logic is EXPTIME-complete as for the underlying ALC.
Tasks
Published 2018-11-06
URL https://arxiv.org/abs/1811.02366v4
PDF https://arxiv.org/pdf/1811.02366v4.pdf
PWC https://paperswithcode.com/paper/a-description-logic-framework-for-commonsense
Repo
Framework

A Decomposition-Based Many-Objective Evolutionary Algorithm with Local Iterative Update

Title A Decomposition-Based Many-Objective Evolutionary Algorithm with Local Iterative Update
Authors Yingyu Zhang, Bing Zeng
Abstract Existing studies have shown that the conventional multi-objective evolutionary algorithms (MOEAs) based on decomposition may lose the population diversity when solving some many-objective optimization problems. In this paper, a simple decomposition-based MOEA with local iterative update (LIU) is proposed. The LIU strategy has two features that are expected to drive the population to approximate the Pareto Front with good distribution. One is that only the worst solution in the current neighborhood is swapped out by the newly generated offspring, preventing the population from being occupied by copies of a few individuals. The other is that its iterative process helps to assign better solutions to subproblems, which is beneficial to make full use of the similarity of solutions to neighboring subproblems and explore local areas in the search space. In addition, the time complexity of the proposed algorithm is the same as that of MOEA/D, and lower than that of other known MOEAs, since it considers only individuals within the current neighborhood at each update. The algorithm is compared with several of the best MOEAs on problems chosen from two famous test suites DTLZ and WFG. Experimental results demonstrate that only a handful of running instances of the algorithm on DTLZ4 lose their population diversity. What’s more, the algorithm wins in most of the test instances in terms of both running time and solution quality, indicating that it is very effective in solving MaOPs.
Tasks
Published 2018-06-27
URL http://arxiv.org/abs/1806.10950v1
PDF http://arxiv.org/pdf/1806.10950v1.pdf
PWC https://paperswithcode.com/paper/a-decomposition-based-many-objective
Repo
Framework

Analysis of adversarial attacks against CNN-based image forgery detectors

Title Analysis of adversarial attacks against CNN-based image forgery detectors
Authors Diego Gragnaniello, Francesco Marra, Giovanni Poggi, Luisa Verdoliva
Abstract With the ubiquitous diffusion of social networks, images are becoming a dominant and powerful communication channel. Not surprisingly, they are also increasingly subject to manipulations aimed at distorting information and spreading fake news. In recent years, the scientific community has devoted major efforts to contrast this menace, and many image forgery detectors have been proposed. Currently, due to the success of deep learning in many multimedia processing tasks, there is high interest towards CNN-based detectors, and early results are already very promising. Recent studies in computer vision, however, have shown CNNs to be highly vulnerable to adversarial attacks, small perturbations of the input data which drive the network towards erroneous classification. In this paper we analyze the vulnerability of CNN-based image forensics methods to adversarial attacks, considering several detectors and several types of attack, and testing performance on a wide range of common manipulations, both easily and hardly detectable.
Tasks
Published 2018-08-25
URL http://arxiv.org/abs/1808.08426v1
PDF http://arxiv.org/pdf/1808.08426v1.pdf
PWC https://paperswithcode.com/paper/analysis-of-adversarial-attacks-against-cnn
Repo
Framework

Sea surface temperature prediction and reconstruction using patch-level neural network representations

Title Sea surface temperature prediction and reconstruction using patch-level neural network representations
Authors Said Ouala, Cedric Herzet, Ronan Fablet
Abstract The forecasting and reconstruction of ocean and atmosphere dynamics from satellite observation time series are key challenges. While model-driven representations remain the classic approaches, data-driven representations become more and more appealing to benefit from available large-scale observation and simulation datasets. In this work we investigate the relevance of recently introduced bilinear residual neural network representations, which mimic numerical integration schemes such as Runge-Kutta, for the forecasting and assimilation of geophysical fields from satellite-derived remote sensing data. As a case-study, we consider satellite-derived Sea Surface Temperature time series off South Africa, which involves intense and complex upper ocean dynamics. Our numerical experiments demonstrate that the proposed patch-level neural-network-based representations outperform other data-driven models, including analog schemes, both in terms of forecasting and missing data interpolation performance with a relative gain up to 50% for highly dynamic areas.
Tasks Time Series
Published 2018-06-01
URL http://arxiv.org/abs/1806.00144v1
PDF http://arxiv.org/pdf/1806.00144v1.pdf
PWC https://paperswithcode.com/paper/sea-surface-temperature-prediction-and
Repo
Framework

Unsupervised Semantic Deep Hashing

Title Unsupervised Semantic Deep Hashing
Authors Sheng Jin
Abstract In recent years, deep hashing methods have been proved to be efficient since it employs convolutional neural network to learn features and hashing codes simultaneously. However, these methods are mostly supervised. In real-world application, it is a time-consuming and overloaded task for annotating a large number of images. In this paper, we propose a novel unsupervised deep hashing method for large-scale image retrieval. Our method, namely unsupervised semantic deep hashing (\textbf{USDH}), uses semantic information preserved in the CNN feature layer to guide the training of network. We enforce four criteria on hashing codes learning based on VGG-19 model: 1) preserving relevant information of feature space in hashing space; 2) minimizing quantization loss between binary-like codes and hashing codes; 3) improving the usage of each bit in hashing codes by using maximum information entropy, and 4) invariant to image rotation. Extensive experiments on CIFAR-10, NUSWIDE have demonstrated that \textbf{USDH} outperforms several state-of-the-art unsupervised hashing methods for image retrieval. We also conduct experiments on Oxford 17 datasets for fine-grained classification to verify its efficiency for other computer vision tasks.
Tasks Image Retrieval, Quantization
Published 2018-03-19
URL http://arxiv.org/abs/1803.06911v1
PDF http://arxiv.org/pdf/1803.06911v1.pdf
PWC https://paperswithcode.com/paper/unsupervised-semantic-deep-hashing
Repo
Framework

Recent Advances in Autoencoder-Based Representation Learning

Title Recent Advances in Autoencoder-Based Representation Learning
Authors Michael Tschannen, Olivier Bachem, Mario Lucic
Abstract Learning useful representations with little or no supervision is a key challenge in artificial intelligence. We provide an in-depth review of recent advances in representation learning with a focus on autoencoder-based models. To organize these results we make use of meta-priors believed useful for downstream tasks, such as disentanglement and hierarchical organization of features. In particular, we uncover three main mechanisms to enforce such properties, namely (i) regularizing the (approximate or aggregate) posterior distribution, (ii) factorizing the encoding and decoding distribution, or (iii) introducing a structured prior distribution. While there are some promising results, implicit or explicit supervision remains a key enabler and all current methods use strong inductive biases and modeling assumptions. Finally, we provide an analysis of autoencoder-based representation learning through the lens of rate-distortion theory and identify a clear tradeoff between the amount of prior knowledge available about the downstream tasks, and how useful the representation is for this task.
Tasks Representation Learning
Published 2018-12-12
URL http://arxiv.org/abs/1812.05069v1
PDF http://arxiv.org/pdf/1812.05069v1.pdf
PWC https://paperswithcode.com/paper/recent-advances-in-autoencoder-based
Repo
Framework

SWAG: A Large-Scale Adversarial Dataset for Grounded Commonsense Inference

Title SWAG: A Large-Scale Adversarial Dataset for Grounded Commonsense Inference
Authors Rowan Zellers, Yonatan Bisk, Roy Schwartz, Yejin Choi
Abstract Given a partial description like “she opened the hood of the car,” humans can reason about the situation and anticipate what might come next (“then, she examined the engine”). In this paper, we introduce the task of grounded commonsense inference, unifying natural language inference and commonsense reasoning. We present SWAG, a new dataset with 113k multiple choice questions about a rich spectrum of grounded situations. To address the recurring challenges of the annotation artifacts and human biases found in many existing datasets, we propose Adversarial Filtering (AF), a novel procedure that constructs a de-biased dataset by iteratively training an ensemble of stylistic classifiers, and using them to filter the data. To account for the aggressive adversarial filtering, we use state-of-the-art language models to massively oversample a diverse set of potential counterfactuals. Empirical results demonstrate that while humans can solve the resulting inference problems with high accuracy (88%), various competitive models struggle on our task. We provide comprehensive analysis that indicates significant opportunities for future research.
Tasks Common Sense Reasoning, Natural Language Inference, Question Answering
Published 2018-08-16
URL http://arxiv.org/abs/1808.05326v1
PDF http://arxiv.org/pdf/1808.05326v1.pdf
PWC https://paperswithcode.com/paper/swag-a-large-scale-adversarial-dataset-for
Repo
Framework

ProdSumNet: reducing model parameters in deep neural networks via product-of-sums matrix decompositions

Title ProdSumNet: reducing model parameters in deep neural networks via product-of-sums matrix decompositions
Authors Chai Wah Wu
Abstract We consider a general framework for reducing the number of trainable model parameters in deep learning networks by decomposing linear operators as a product of sums of simpler linear operators. Recently proposed deep learning architectures such as CNN, KFC, Dilated CNN, etc. are all subsumed in this framework and we illustrate other types of neural network architectures within this framework. We show that good accuracy on MNIST and Fashion MNIST can be obtained using a relatively small number of trainable parameters. In addition, since implementation of the convolutional layer is resource-heavy, we consider an approach in the transform domain that obviates the need for convolutional layers. One of the advantages of this general framework over prior approaches is that the number of trainable parameters is not fixed and can be varied arbitrarily. In particular, we illustrate the tradeoff of varying the number of trainable variables and the corresponding error rate. As an example, by using this decomposition on a reference CNN architecture for MNIST with over 3x10^6 trainable parameters, we are able to obtain an accuracy of 98.44% using only 3554 trainable parameters.
Tasks
Published 2018-09-06
URL https://arxiv.org/abs/1809.02209v2
PDF https://arxiv.org/pdf/1809.02209v2.pdf
PWC https://paperswithcode.com/paper/prodsumnet-reducing-model-parameters-in-deep
Repo
Framework

DeepDPM: Dynamic Population Mapping via Deep Neural Network

Title DeepDPM: Dynamic Population Mapping via Deep Neural Network
Authors Zefang Zong, Jie Feng, Kechun Liu, Hongzhi Shi, Yong Li
Abstract Dynamic high resolution data on human population distribution is of great importance for a wide spectrum of activities and real-life applications, but is too difficult and expensive to obtain directly. Therefore, generating fine-scaled population distributions from coarse population data is of great significance. However, there are three major challenges: 1) the complexity in spatial relations between high and low resolution population; 2) the dependence of population distributions on other external information; 3) the difficulty in retrieving temporal distribution patterns. In this paper, we first propose the idea to generate dynamic population distributions in full-time series, then we design dynamic population mapping via deep neural network(DeepDPM), a model that describes both spatial and temporal patterns using coarse data and point of interest information. In DeepDPM, we utilize super-resolution convolutional neural network(SRCNN) based model to directly map coarse data into higher resolution data, and a time-embedded long short-term memory model to effectively capture the periodicity nature to smooth the finer-scaled results from the previous static SRCNN model. We perform extensive experiments on a real-life mobile dataset collected from Shanghai. Our results demonstrate that DeepDPM outperforms previous state-of-the-art methods and a suite of frequent data-mining approaches. Moreover, DeepDPM breaks through the limitation from previous works in time dimension so that dynamic predictions in all-day time slots can be obtained.
Tasks Population Mapping, Super-Resolution, Time Series
Published 2018-10-25
URL http://arxiv.org/abs/1811.02644v2
PDF http://arxiv.org/pdf/1811.02644v2.pdf
PWC https://paperswithcode.com/paper/deepdpm-dynamic-population-mapping-via-deep
Repo
Framework

Multi-modal Disease Classification in Incomplete Datasets Using Geometric Matrix Completion

Title Multi-modal Disease Classification in Incomplete Datasets Using Geometric Matrix Completion
Authors Gerome Vivar, Andreas Zwergal, Nassir Navab, Seyed-Ahmad Ahmadi
Abstract In large population-based studies and in clinical routine, tasks like disease diagnosis and progression prediction are inherently based on a rich set of multi-modal data, including imaging and other sensor data, clinical scores, phenotypes, labels and demographics. However, missing features, rater bias and inaccurate measurements are typical ailments of real-life medical datasets. Recently, it has been shown that deep learning with graph convolution neural networks (GCN) can outperform traditional machine learning in disease classification, but missing features remain an open problem. In this work, we follow up on the idea of modeling multi-modal disease classification as a matrix completion problem, with simultaneous classification and non-linear imputation of features. Compared to methods before, we arrange subjects in a graph-structure and solve classification through geometric matrix completion, which simulates a heat diffusion process that is learned and solved with a recurrent neural network. We demonstrate the potential of this method on the ADNI-based TADPOLE dataset and on the task of predicting the transition from MCI to Alzheimer’s disease. With an AUC of 0.950 and classification accuracy of 87%, our approach outperforms standard linear and non-linear classifiers, as well as several state-of-the-art results in related literature, including a recently proposed GCN-based approach.
Tasks Imputation, Matrix Completion
Published 2018-03-30
URL http://arxiv.org/abs/1803.11550v1
PDF http://arxiv.org/pdf/1803.11550v1.pdf
PWC https://paperswithcode.com/paper/multi-modal-disease-classification-in
Repo
Framework

Stein Variational Gradient Descent as Moment Matching

Title Stein Variational Gradient Descent as Moment Matching
Authors Qiang Liu, Dilin Wang
Abstract Stein variational gradient descent (SVGD) is a non-parametric inference algorithm that evolves a set of particles to fit a given distribution of interest. We analyze the non-asymptotic properties of SVGD, showing that there exists a set of functions, which we call the Stein matching set, whose expectations are exactly estimated by any set of particles that satisfies the fixed point equation of SVGD. This set is the image of Stein operator applied on the feature maps of the positive definite kernel used in SVGD. Our results provide a theoretical framework for analyzing the properties of SVGD with different kernels, shedding insight into optimal kernel choice. In particular, we show that SVGD with linear kernels yields exact estimation of means and variances on Gaussian distributions, while random Fourier features enable probabilistic bounds for distributional approximation. Our results offer a refreshing view of the classical inference problem as fitting Stein’s identity or solving the Stein equation, which may motivate more efficient algorithms.
Tasks
Published 2018-10-27
URL http://arxiv.org/abs/1810.11693v1
PDF http://arxiv.org/pdf/1810.11693v1.pdf
PWC https://paperswithcode.com/paper/stein-variational-gradient-descent-as-moment
Repo
Framework

Floyd-Warshall Reinforcement Learning: Learning from Past Experiences to Reach New Goals

Title Floyd-Warshall Reinforcement Learning: Learning from Past Experiences to Reach New Goals
Authors Vikas Dhiman, Shurjo Banerjee, Jeffrey M. Siskind, Jason J. Corso
Abstract Consider mutli-goal tasks that involve static environments and dynamic goals. Examples of such tasks, such as goal-directed navigation and pick-and-place in robotics, abound. Two types of Reinforcement Learning (RL) algorithms are used for such tasks: model-free or model-based. Each of these approaches has limitations. Model-free RL struggles to transfer learned information when the goal location changes, but achieves high asymptotic accuracy in single goal tasks. Model-based RL can transfer learned information to new goal locations by retaining the explicitly learned state-dynamics, but is limited by the fact that small errors in modelling these dynamics accumulate over long-term planning. In this work, we improve upon the limitations of model-free RL in multi-goal domains. We do this by adapting the Floyd-Warshall algorithm for RL and call the adaptation Floyd-Warshall RL (FWRL). The proposed algorithm learns a goal-conditioned action-value function by constraining the value of the optimal path between any two states to be greater than or equal to the value of paths via intermediary states. Experimentally, we show that FWRL is more sample-efficient and learns higher reward strategies in multi-goal tasks as compared to Q-learning, model-based RL and other relevant baselines in a tabular domain.
Tasks Q-Learning
Published 2018-09-25
URL http://arxiv.org/abs/1809.09318v4
PDF http://arxiv.org/pdf/1809.09318v4.pdf
PWC https://paperswithcode.com/paper/floyd-warshall-reinforcement-learning
Repo
Framework

All-Optical Machine Learning Using Diffractive Deep Neural Networks

Title All-Optical Machine Learning Using Diffractive Deep Neural Networks
Authors Xing Lin, Yair Rivenson, Nezih T. Yardimci, Muhammed Veli, Mona Jarrahi, Aydogan Ozcan
Abstract We introduce an all-optical Diffractive Deep Neural Network (D2NN) architecture that can learn to implement various functions after deep learning-based design of passive diffractive layers that work collectively. We experimentally demonstrated the success of this framework by creating 3D-printed D2NNs that learned to implement handwritten digit classification and the function of an imaging lens at terahertz spectrum. With the existing plethora of 3D-printing and other lithographic fabrication methods as well as spatial-light-modulators, this all-optical deep learning framework can perform, at the speed of light, various complex functions that computer-based neural networks can implement, and will find applications in all-optical image analysis, feature detection and object classification, also enabling new camera designs and optical components that can learn to perform unique tasks using D2NNs.
Tasks Object Classification
Published 2018-04-14
URL http://arxiv.org/abs/1804.08711v2
PDF http://arxiv.org/pdf/1804.08711v2.pdf
PWC https://paperswithcode.com/paper/all-optical-machine-learning-using
Repo
Framework

Transductive Boltzmann Machines

Title Transductive Boltzmann Machines
Authors Mahito Sugiyama, Koji Tsuda, Hiroyuki Nakahara
Abstract We present transductive Boltzmann machines (TBMs), which firstly achieve transductive learning of the Gibbs distribution. While exact learning of the Gibbs distribution is impossible by the family of existing Boltzmann machines due to combinatorial explosion of the sample space, TBMs overcome the problem by adaptively constructing the minimum required sample space from data to avoid unnecessary generalization. We theoretically provide bias-variance decomposition of the KL divergence in TBMs to analyze its learnability, and empirically demonstrate that TBMs are superior to the fully visible Boltzmann machines and popularly used restricted Boltzmann machines in terms of efficiency and effectiveness.
Tasks
Published 2018-05-21
URL http://arxiv.org/abs/1805.07938v1
PDF http://arxiv.org/pdf/1805.07938v1.pdf
PWC https://paperswithcode.com/paper/transductive-boltzmann-machines
Repo
Framework

Robust Optimization through Neuroevolution

Title Robust Optimization through Neuroevolution
Authors Paolo Pagliuca, Stefano Nolfi
Abstract We propose a method for evolving solutions that are robust with respect to variations of the environmental conditions (i.e. that can operate effectively in new conditions immediately, without the need to adapt to variations). The obtained results show how the method proposed is effective and computational tractable. It permits to improve performance on an extended version of the double-pole balancing problem, to outperform the best available human-designed controllers on a car racing problem, and to generate rather effective solutions for a swarm robotic problem. The comparison of different algorithms indicates that the CMA-ES and xNES methods, that operate by optimizing a distribution of parameters, represent the best options for the evolution of robust neural network controllers.
Tasks Car Racing
Published 2018-10-02
URL http://arxiv.org/abs/1810.01125v1
PDF http://arxiv.org/pdf/1810.01125v1.pdf
PWC https://paperswithcode.com/paper/robust-optimization-through-neuroevolution
Repo
Framework
comments powered by Disqus