February 1, 2020

2902 words 14 mins read

Paper Group AWR 142

Paper Group AWR 142

Privacy-Preserving Obfuscation of Critical Infrastructure Networks. Addressing Marketing Bias in Product Recommendations. Multi-view Information-theoretic Co-clustering for Co-occurrence Data. D2-Net: A Trainable CNN for Joint Detection and Description of Local Features. The Second Conversational Intelligence Challenge (ConvAI2). Towards Characteri …

Privacy-Preserving Obfuscation of Critical Infrastructure Networks

Title Privacy-Preserving Obfuscation of Critical Infrastructure Networks
Authors Ferdinando Fioretto, Terrence W. K. Mak, Pascal Van Hentenryck
Abstract The paper studies how to release data about a critical infrastructure network (e.g., the power network or a transportation network) without disclosing sensitive information that can be exploited by malevolent agents, while preserving the realism of the network. It proposes a novel obfuscation mechanism that combines several privacy-preserving building blocks with a bi-level optimization model to significantly improve accuracy. The obfuscation is evaluated for both realism and privacy properties on real energy and transportation networks. Experimental results show the obfuscation mechanism substantially reduces the potential damage of an attack exploiting the released data to harm the real network.
Tasks
Published 2019-05-23
URL https://arxiv.org/abs/1905.09778v2
PDF https://arxiv.org/pdf/1905.09778v2.pdf
PWC https://paperswithcode.com/paper/privacy-preserving-obfuscation-of-critical
Repo https://github.com/Chrisackerman1/Privacy-Preserving-Obfuscation-of-Critical-Infrastructure-Networks
Framework none

Addressing Marketing Bias in Product Recommendations

Title Addressing Marketing Bias in Product Recommendations
Authors Mengting Wan, Jianmo Ni, Rishabh Misra, Julian McAuley
Abstract Modern collaborative filtering algorithms seek to provide personalized product recommendations by uncovering patterns in consumer-product interactions. However, these interactions can be biased by how the product is marketed, for example due to the selection of a particular human model in a product image. These correlations may result in the underrepresentation of particular niche markets in the interaction data; for example, a female user who would potentially like motorcycle products may be less likely to interact with them if they are promoted using stereotypically ‘male’ images. In this paper, we first investigate this correlation between users’ interaction feedback and products’ marketing images on two real-world e-commerce datasets. We further examine the response of several standard collaborative filtering algorithms to the distribution of consumer-product market segments in the input interaction data, revealing that marketing strategy can be a source of bias for modern recommender systems. In order to protect recommendation performance on underrepresented market segments, we develop a framework to address this potential marketing bias. Quantitative results demonstrate that the proposed approach significantly improves the recommendation fairness across different market segments, with a negligible loss (or better) recommendation accuracy.
Tasks Recommendation Systems
Published 2019-12-04
URL https://arxiv.org/abs/1912.01799v1
PDF https://arxiv.org/pdf/1912.01799v1.pdf
PWC https://paperswithcode.com/paper/addressing-marketing-bias-in-product
Repo https://github.com/MengtingWan/marketBias
Framework tf

Multi-view Information-theoretic Co-clustering for Co-occurrence Data

Title Multi-view Information-theoretic Co-clustering for Co-occurrence Data
Authors Peng Xu, Zhaohong Deng, Kup-Sze Choi, Longbing Cao, Shitong Wang
Abstract Multi-view clustering has received much attention recently. Most of the existing multi-view clustering methods only focus on one-sided clustering. As the co-occurring data elements involve the counts of sample-feature co-occurrences, it is more efficient to conduct two-sided clustering along the samples and features simultaneously. To take advantage of two-sided clustering for the co-occurrences in the scene of multi-view clustering, a two-sided multi-view clustering method is proposed, i.e., multi-view information-theoretic co-clustering (MV-ITCC). The proposed method realizes two-sided clustering for co-occurring multi-view data under the formulation of information theory. More specifically, it exploits the agreement and disagreement among views by sharing a common clustering results along the sample dimension and keeping the clustering results of each view specific along the feature dimension. In addition, the mechanism of maximum entropy is also adopted to control the importance of different views, which can give a right balance in leveraging the agreement and disagreement. Extensive experiments are conducted on text and image multi-view datasets. The results clearly demonstrate the superiority of the proposed method.
Tasks
Published 2019-05-25
URL https://arxiv.org/abs/1905.10594v1
PDF https://arxiv.org/pdf/1905.10594v1.pdf
PWC https://paperswithcode.com/paper/multi-view-information-theoretic-co
Repo https://github.com/DallasBuyer/MVITCC
Framework none

D2-Net: A Trainable CNN for Joint Detection and Description of Local Features

Title D2-Net: A Trainable CNN for Joint Detection and Description of Local Features
Authors Mihai Dusmanu, Ignacio Rocco, Tomas Pajdla, Marc Pollefeys, Josef Sivic, Akihiko Torii, Torsten Sattler
Abstract In this work we address the problem of finding reliable pixel-level correspondences under difficult imaging conditions. We propose an approach where a single convolutional neural network plays a dual role: It is simultaneously a dense feature descriptor and a feature detector. By postponing the detection to a later stage, the obtained keypoints are more stable than their traditional counterparts based on early detection of low-level structures. We show that this model can be trained using pixel correspondences extracted from readily available large-scale SfM reconstructions, without any further annotations. The proposed method obtains state-of-the-art performance on both the difficult Aachen Day-Night localization dataset and the InLoc indoor localization benchmark, as well as competitive performance on other benchmarks for image matching and 3D reconstruction.
Tasks 3D Reconstruction
Published 2019-05-09
URL https://arxiv.org/abs/1905.03561v1
PDF https://arxiv.org/pdf/1905.03561v1.pdf
PWC https://paperswithcode.com/paper/190503561
Repo https://github.com/mihaidusmanu/d2-net
Framework pytorch

The Second Conversational Intelligence Challenge (ConvAI2)

Title The Second Conversational Intelligence Challenge (ConvAI2)
Authors Emily Dinan, Varvara Logacheva, Valentin Malykh, Alexander Miller, Kurt Shuster, Jack Urbanek, Douwe Kiela, Arthur Szlam, Iulian Serban, Ryan Lowe, Shrimai Prabhumoye, Alan W Black, Alexander Rudnicky, Jason Williams, Joelle Pineau, Mikhail Burtsev, Jason Weston
Abstract We describe the setting and results of the ConvAI2 NeurIPS competition that aims to further the state-of-the-art in open-domain chatbots. Some key takeaways from the competition are: (i) pretrained Transformer variants are currently the best performing models on this task, (ii) but to improve performance on multi-turn conversations with humans, future systems must go beyond single word metrics like perplexity to measure the performance across sequences of utterances (conversations) – in terms of repetition, consistency and balance of dialogue acts (e.g. how many questions asked vs. answered).
Tasks
Published 2019-01-31
URL http://arxiv.org/abs/1902.00098v1
PDF http://arxiv.org/pdf/1902.00098v1.pdf
PWC https://paperswithcode.com/paper/the-second-conversational-intelligence
Repo https://github.com/DeepPavlov/convai
Framework pytorch

Towards Characterizing and Limiting Information Exposure in DNN Layers

Title Towards Characterizing and Limiting Information Exposure in DNN Layers
Authors Fan Mo, Ali Shahin Shamsabadi, Kleomenis Katevas, Andrea Cavallaro, Hamed Haddadi
Abstract Pre-trained Deep Neural Network (DNN) models are increasingly used in smartphones and other user devices to enable prediction services, leading to potential disclosures of (sensitive) information from training data captured inside these models. Based on the concept of generalization error, we propose a framework to measure the amount of sensitive information memorized in each layer of a DNN. Our results show that, when considered individually, the last layers encode a larger amount of information from the training data compared to the first layers. We find that, while the neuron of convolutional layers can expose more (sensitive) information than that of fully connected layers, the same DNN architecture trained with different datasets has similar exposure per layer. We evaluate an architecture to protect the most sensitive layers within the memory limits of Trusted Execution Environment (TEE) against potential white-box membership inference attacks without the significant computational overhead.
Tasks
Published 2019-07-13
URL https://arxiv.org/abs/1907.06034v1
PDF https://arxiv.org/pdf/1907.06034v1.pdf
PWC https://paperswithcode.com/paper/towards-characterizing-and-limiting
Repo https://github.com/mofanv/darknetp
Framework none

Transfer Learning by Modeling a Distribution over Policies

Title Transfer Learning by Modeling a Distribution over Policies
Authors Disha Shrivastava, Eeshan Gunesh Dhekane, Riashat Islam
Abstract Exploration and adaptation to new tasks in a transfer learning setup is a central challenge in reinforcement learning. In this work, we build on the idea of modeling a distribution over policies in a Bayesian deep reinforcement learning setup to propose a transfer strategy. Recent works have shown to induce diversity in the learned policies by maximizing the entropy of a distribution of policies (Bachman et al., 2018; Garnelo et al., 2018) and thus, we postulate that our proposed approach leads to faster exploration resulting in improved transfer learning. We support our hypothesis by demonstrating favorable experimental results on a variety of settings on fully-observable GridWorld and partially observable MiniGrid (Chevalier-Boisvert et al., 2018) environments.
Tasks Transfer Learning
Published 2019-06-09
URL https://arxiv.org/abs/1906.03574v1
PDF https://arxiv.org/pdf/1906.03574v1.pdf
PWC https://paperswithcode.com/paper/transfer-learning-by-modeling-a-distribution
Repo https://github.com/maximecb/gym-minigrid
Framework pytorch

Convergent Policy Optimization for Safe Reinforcement Learning

Title Convergent Policy Optimization for Safe Reinforcement Learning
Authors Ming Yu, Zhuoran Yang, Mladen Kolar, Zhaoran Wang
Abstract We study the safe reinforcement learning problem with nonlinear function approximation, where policy optimization is formulated as a constrained optimization problem with both the objective and the constraint being nonconvex functions. For such a problem, we construct a sequence of surrogate convex constrained optimization problems by replacing the nonconvex functions locally with convex quadratic functions obtained from policy gradient estimators. We prove that the solutions to these surrogate problems converge to a stationary point of the original nonconvex problem. Furthermore, to extend our theoretical results, we apply our algorithm to examples of optimal control and multi-agent reinforcement learning with safety constraints.
Tasks Multi-agent Reinforcement Learning
Published 2019-10-26
URL https://arxiv.org/abs/1910.12156v1
PDF https://arxiv.org/pdf/1910.12156v1.pdf
PWC https://paperswithcode.com/paper/convergent-policy-optimization-for-safe
Repo https://github.com/ming93/Safe_reinforcement_learning
Framework none

An Interactive Multi-Task Learning Network for End-to-End Aspect-Based Sentiment Analysis

Title An Interactive Multi-Task Learning Network for End-to-End Aspect-Based Sentiment Analysis
Authors Ruidan He, Wee Sun Lee, Hwee Tou Ng, Daniel Dahlmeier
Abstract Aspect-based sentiment analysis produces a list of aspect terms and their corresponding sentiments for a natural language sentence. This task is usually done in a pipeline manner, with aspect term extraction performed first, followed by sentiment predictions toward the extracted aspect terms. While easier to develop, such an approach does not fully exploit joint information from the two subtasks and does not use all available sources of training information that might be helpful, such as document-level labeled sentiment corpus. In this paper, we propose an interactive multi-task learning network (IMN) which is able to jointly learn multiple related tasks simultaneously at both the token level as well as the document level. Unlike conventional multi-task learning methods that rely on learning common features for the different tasks, IMN introduces a message passing architecture where information is iteratively passed to different tasks through a shared set of latent variables. Experimental results demonstrate superior performance of the proposed method against multiple baselines on three benchmark datasets.
Tasks Aspect-Based Sentiment Analysis, Multi-Task Learning, Sentiment Analysis
Published 2019-06-17
URL https://arxiv.org/abs/1906.06906v1
PDF https://arxiv.org/pdf/1906.06906v1.pdf
PWC https://paperswithcode.com/paper/an-interactive-multi-task-learning-network
Repo https://github.com/ruidan/IMN-E2E-ABSA
Framework tf

Generative Adversarial Minority Oversampling

Title Generative Adversarial Minority Oversampling
Authors Sankha Subhra Mullick, Shounak Datta, Swagatam Das
Abstract Class imbalance is a long-standing problem relevant to a number of real-world applications of deep learning. Oversampling techniques, which are effective for handling class imbalance in classical learning systems, can not be directly applied to end-to-end deep learning systems. We propose a three-player adversarial game between a convex generator, a multi-class classifier network, and a real/fake discriminator to perform oversampling in deep learning systems. The convex generator generates new samples from the minority classes as convex combinations of existing instances, aiming to fool both the discriminator as well as the classifier into misclassifying the generated samples. Consequently, the artificial samples are generated at critical locations near the peripheries of the classes. This, in turn, adjusts the classifier induced boundaries in a way which is more likely to reduce misclassification from the minority classes. Extensive experiments on multiple class imbalanced image datasets establish the efficacy of our proposal.
Tasks
Published 2019-03-22
URL http://arxiv.org/abs/1903.09730v2
PDF http://arxiv.org/pdf/1903.09730v2.pdf
PWC https://paperswithcode.com/paper/generative-adversarial-minority-oversampling
Repo https://github.com/SankhaSubhra/GAMO
Framework none

Regularized Learning for Domain Adaptation under Label Shifts

Title Regularized Learning for Domain Adaptation under Label Shifts
Authors Kamyar Azizzadenesheli, Anqi Liu, Fanny Yang, Animashree Anandkumar
Abstract We propose Regularized Learning under Label shifts (RLLS), a principled and a practical domain-adaptation algorithm to correct for shifts in the label distribution between a source and a target domain. We first estimate importance weights using labeled source data and unlabeled target data, and then train a classifier on the weighted source samples. We derive a generalization bound for the classifier on the target domain which is independent of the (ambient) data dimensions, and instead only depends on the complexity of the function class. To the best of our knowledge, this is the first generalization bound for the label-shift problem where the labels in the target domain are not available. Based on this bound, we propose a regularized estimator for the small-sample regime which accounts for the uncertainty in the estimated weights. Experiments on the CIFAR-10 and MNIST datasets show that RLLS improves classification accuracy, especially in the low sample and large-shift regimes, compared to previous methods.
Tasks Domain Adaptation
Published 2019-03-22
URL http://arxiv.org/abs/1903.09734v1
PDF http://arxiv.org/pdf/1903.09734v1.pdf
PWC https://paperswithcode.com/paper/regularized-learning-for-domain-adaptation-1
Repo https://github.com/Angela0428/labelshift
Framework pytorch

Training Deep SLAM on Single Frames

Title Training Deep SLAM on Single Frames
Authors Igor Slinko, Anna Vorontsova, Dmitry Zhukov, Olga Barinova, Anton Konushin
Abstract Learning-based visual odometry and SLAM methods demonstrate a steady improvement over past years. However, collecting ground truth poses to train these methods is difficult and expensive. This could be resolved by training in an unsupervised mode, but there is still a large gap between performance of unsupervised and supervised methods. In this work, we focus on generating synthetic data for deep learning-based visual odometry and SLAM methods that take optical flow as an input. We produce training data in a form of optical flow that corresponds to arbitrary camera movement between a real frame and a virtual frame. For synthesizing data we use depth maps either produced by a depth sensor or estimated from stereo pair. We train visual odometry model on synthetic data and do not use ground truth poses hence this model can be considered unsupervised. Also it can be classified as monocular as we do not use depth maps on inference. We also propose a simple way to convert any visual odometry model into a SLAM method based on frame matching and graph optimization. We demonstrate that both the synthetically-trained visual odometry model and the proposed SLAM method build upon this model yields state-of-the-art results among unsupervised methods on KITTI dataset and shows promising results on a challenging EuRoC dataset.
Tasks Optical Flow Estimation, Visual Odometry
Published 2019-12-11
URL https://arxiv.org/abs/1912.05405v1
PDF https://arxiv.org/pdf/1912.05405v1.pdf
PWC https://paperswithcode.com/paper/training-deep-slam-on-single-frames
Repo https://github.com/saic-vul/odometry
Framework none

MadMiner: Machine learning-based inference for particle physics

Title MadMiner: Machine learning-based inference for particle physics
Authors Johann Brehmer, Felix Kling, Irina Espejo, Kyle Cranmer
Abstract Precision measurements at the LHC often require analyzing high-dimensional event data for subtle kinematic signatures, which is challenging for established analysis methods. Recently, a powerful family of multivariate inference techniques that leverage both matrix element information and machine learning has been developed. This approach neither requires the reduction of high-dimensional data to summary statistics nor any simplifications to the underlying physics or detector response. In this paper we introduce MadMiner, a Python module that streamlines the steps involved in this procedure. Wrapping around MadGraph5_aMC and Pythia 8, it supports almost any physics process and model. To aid phenomenological studies, the tool also wraps around Delphes 3, though it is extendable to a full Geant4-based detector simulation. We demonstrate the use of MadMiner in an example analysis of dimension-six operators in ttH production, finding that the new techniques substantially increase the sensitivity to new physics.
Tasks
Published 2019-07-24
URL https://arxiv.org/abs/1907.10621v2
PDF https://arxiv.org/pdf/1907.10621v2.pdf
PWC https://paperswithcode.com/paper/madminer-machine-learning-based-inference-for
Repo https://github.com/diana-hep/madminer
Framework pytorch

Exploiting GAN Internal Capacity for High-Quality Reconstruction of Natural Images

Title Exploiting GAN Internal Capacity for High-Quality Reconstruction of Natural Images
Authors Marcos Pividori, Guillermo L. Grinblat, Lucas C. Uzal
Abstract Generative Adversarial Networks (GAN) have demonstrated impressive results in modeling the distribution of natural images, learning latent representations that capture semantic variations in an unsupervised basis. Beyond the generation of novel samples, it is of special interest to exploit the ability of the GAN generator to model the natural image manifold and hence generate credible changes when manipulating images. However, this line of work is conditioned by the quality of the reconstruction. Until now, only inversion to the latent space has been considered, we propose to exploit the representation in intermediate layers of the generator, and we show that this leads to increased capacity. In particular, we observe that the representation after the first dense layer, present in all state-of-the-art GAN models, is expressive enough to represent natural images with high visual fidelity. It is possible to interpolate around these images obtaining a sequence of new plausible synthetic images that cannot be generated from the latent space. Finally, as an example of potential applications that arise from this inversion mechanism, we show preliminary results in exploiting the learned representation in the attention map of the generator to obtain an unsupervised segmentation of natural images.
Tasks
Published 2019-10-26
URL https://arxiv.org/abs/1911.05630v1
PDF https://arxiv.org/pdf/1911.05630v1.pdf
PWC https://paperswithcode.com/paper/exploiting-gan-internal-capacity-for-high
Repo https://github.com/CIFASIS/exploiting-gan-internal-capacity
Framework tf

Efficient Evolutionary Methods for Game Agent Optimisation: Model-Based is Best

Title Efficient Evolutionary Methods for Game Agent Optimisation: Model-Based is Best
Authors Simon M. Lucas, Jialin Liu, Ivan Bravi, Raluca D. Gaina, John Woodward, Vanessa Volz, Diego Perez-Liebana
Abstract This paper introduces a simple and fast variant of Planet Wars as a test-bed for statistical planning based Game AI agents, and for noisy hyper-parameter optimisation. Planet Wars is a real-time strategy game with simple rules but complex game-play. The variant introduced in this paper is designed for speed to enable efficient experimentation, and also for a fixed action space to enable practical inter-operability with General Video Game AI agents. If we treat the game as a win-loss game (which is standard), then this leads to challenging noisy optimisation problems both in tuning agents to play the game, and in tuning game parameters. Here we focus on the problem of tuning an agent, and report results using the recently developed N-Tuple Bandit Evolutionary Algorithm and a number of other optimisers, including Sequential Model-based Algorithm Configuration (SMAC). Results indicate that the N-Tuple Bandit Evolutionary offers competitive performance as well as insight into the effects of combinations of parameter choices.
Tasks
Published 2019-01-03
URL http://arxiv.org/abs/1901.00723v1
PDF http://arxiv.org/pdf/1901.00723v1.pdf
PWC https://paperswithcode.com/paper/efficient-evolutionary-methods-for-game-agent
Repo https://github.com/SimonLucas/ntbea
Framework none
comments powered by Disqus