January 30, 2020

2736 words 13 mins read

Paper Group ANR 346

Paper Group ANR 346

A Comparative Analysis of Knowledge-Intensive and Data-Intensive Semantic Parsers. A Deep Learning approach for Hindi Named Entity Recognition. Experimental Demonstration of Learned Time-Domain Digital Back-Propagation. Experience Sharing Between Cooperative Reinforcement Learning Agents. Perturbed-History Exploration in Stochastic Linear Bandits. …

A Comparative Analysis of Knowledge-Intensive and Data-Intensive Semantic Parsers

Title A Comparative Analysis of Knowledge-Intensive and Data-Intensive Semantic Parsers
Authors Junjie Cao, Zi Lin, Weiwei Sun, Xiaojun Wan
Abstract We present a phenomenon-oriented comparative analysis of the two dominant approaches in task-independent semantic parsing: classic, knowledge-intensive and neural, data-intensive models. To reflect state-of-the-art neural NLP technologies, we introduce a new target structure-centric parser that can produce semantic graphs much more accurately than previous data-driven parsers. We then show that, in spite of comparable performance overall, knowledge- and data-intensive models produce different types of errors, in a way that can be explained by their theoretical properties. This analysis leads to new directions for parser development.
Tasks Semantic Parsing
Published 2019-07-04
URL https://arxiv.org/abs/1907.02298v2
PDF https://arxiv.org/pdf/1907.02298v2.pdf
PWC https://paperswithcode.com/paper/a-comparative-analysis-of-knowledge-intensive
Repo
Framework

A Deep Learning approach for Hindi Named Entity Recognition

Title A Deep Learning approach for Hindi Named Entity Recognition
Authors Bansi Shah, Sunil Kumar Kopparapu
Abstract Named Entity Recognition is one of the most important text processing requirement in many NLP tasks. In this paper we use a deep architecture to accomplish the task of recognizing named entities in a given Hindi text sentence. Bidirectional Long Short Term Memory (BiLSTM) based techniques have been used for NER task in literature. In this paper, we first tune BiLSTM low-resource scenario to work for Hindi NER and propose two enhancements namely (a) de-noising auto-encoder (DAE) LSTM and (b) conditioning LSTM which show improvement in NER task compared to the BiLSTM approach. We use pre-trained word embedding to represent the words in the corpus, and the NER tags of the words are as defined by the used annotated corpora. Experiments have been performed to analyze the performance of different word embeddings and batch sizes which is essential for training deep models.
Tasks Named Entity Recognition, Word Embeddings
Published 2019-11-05
URL https://arxiv.org/abs/1911.01421v1
PDF https://arxiv.org/pdf/1911.01421v1.pdf
PWC https://paperswithcode.com/paper/a-deep-learning-approach-for-hindi-named
Repo
Framework

Experimental Demonstration of Learned Time-Domain Digital Back-Propagation

Title Experimental Demonstration of Learned Time-Domain Digital Back-Propagation
Authors Eric Sillekens, Wenting Yi, Daniel Semrau, Alessandro Ottino, Boris Karanov, Sujie Zhou, Kevin Law, Jack Chen, Domanic Lavery, Lidia Galdino, Polina Bayvel, Robert I. Killey
Abstract We present the first experimental demonstration of learned time-domain digital back-propagation (DBP), in 64-GBd dual-polarization 64-QAM signal transmission over 1014 km. Performance gains were comparable to those obtained with conventional, higher complexity, frequency-domain DBP.
Tasks
Published 2019-12-23
URL https://arxiv.org/abs/1912.12197v1
PDF https://arxiv.org/pdf/1912.12197v1.pdf
PWC https://paperswithcode.com/paper/experimental-demonstration-of-learned-time
Repo
Framework

Experience Sharing Between Cooperative Reinforcement Learning Agents

Title Experience Sharing Between Cooperative Reinforcement Learning Agents
Authors Lucas Oliveira Souza, Gabriel de Oliveira Ramos, Celia Ghedini Ralha
Abstract The idea of experience sharing between cooperative agents naturally emerges from our understanding of how humans learn. Our evolution as a species is tightly linked to the ability to exchange learned knowledge with one another. It follows that experience sharing (ES) between autonomous and independent agents could become the key to accelerate learning in cooperative multiagent settings. We investigate if randomly selecting experiences to share can increase the performance of deep reinforcement learning agents, and propose three new methods for selecting experiences to accelerate the learning process. Firstly, we introduce Focused ES, which prioritizes unexplored regions of the state space. Secondly, we present Prioritized ES, in which temporal-difference error is used as a measure of priority. Finally, we devise Focused Prioritized ES, which combines both previous approaches. The methods are empirically validated in a control problem. While sharing randomly selected experiences between two Deep Q-Network agents shows no improvement over a single agent baseline, we show that the proposed ES methods can successfully outperform the baseline. In particular, the Focused ES accelerates learning by a factor of 2, reducing by 51% the number of episodes required to complete the task.
Tasks
Published 2019-11-06
URL https://arxiv.org/abs/1911.02191v1
PDF https://arxiv.org/pdf/1911.02191v1.pdf
PWC https://paperswithcode.com/paper/experience-sharing-between-cooperative
Repo
Framework

Perturbed-History Exploration in Stochastic Linear Bandits

Title Perturbed-History Exploration in Stochastic Linear Bandits
Authors Branislav Kveton, Csaba Szepesvari, Mohammad Ghavamzadeh, Craig Boutilier
Abstract We propose a new online algorithm for minimizing the cumulative regret in stochastic linear bandits. The key idea is to build a perturbed history, which mixes the history of observed rewards with a pseudo-history of randomly generated i.i.d. pseudo-rewards. Our algorithm, perturbed-history exploration in a linear bandit (LinPHE), estimates a linear model from its perturbed history and pulls the arm with the highest value under that model. We prove a $\tilde{O}(d \sqrt{n})$ gap-free bound on the expected $n$-round regret of LinPHE, where $d$ is the number of features. Our analysis relies on novel concentration and anti-concentration bounds on the weighted sum of Bernoulli random variables. To show the generality of our design, we extend LinPHE to a logistic reward model. We evaluate both algorithms empirically and show that they are practical.
Tasks
Published 2019-03-21
URL http://arxiv.org/abs/1903.09132v1
PDF http://arxiv.org/pdf/1903.09132v1.pdf
PWC https://paperswithcode.com/paper/perturbed-history-exploration-in-stochastic-1
Repo
Framework

Improving Grounded Natural Language Understanding through Human-Robot Dialog

Title Improving Grounded Natural Language Understanding through Human-Robot Dialog
Authors Jesse Thomason, Aishwarya Padmakumar, Jivko Sinapov, Nick Walker, Yuqian Jiang, Harel Yedidsion, Justin Hart, Peter Stone, Raymond J. Mooney
Abstract Natural language understanding for robotics can require substantial domain- and platform-specific engineering. For example, for mobile robots to pick-and-place objects in an environment to satisfy human commands, we can specify the language humans use to issue such commands, and connect concept words like red can to physical object properties. One way to alleviate this engineering for a new domain is to enable robots in human environments to adapt dynamically—continually learning new language constructions and perceptual concepts. In this work, we present an end-to-end pipeline for translating natural language commands to discrete robot actions, and use clarification dialogs to jointly improve language parsing and concept grounding. We train and evaluate this agent in a virtual setting on Amazon Mechanical Turk, and we transfer the learned agent to a physical robot platform to demonstrate it in the real world.
Tasks
Published 2019-03-01
URL http://arxiv.org/abs/1903.00122v1
PDF http://arxiv.org/pdf/1903.00122v1.pdf
PWC https://paperswithcode.com/paper/improving-grounded-natural-language
Repo
Framework

Assessing differentially private deep learning with Membership Inference

Title Assessing differentially private deep learning with Membership Inference
Authors Daniel Bernau, Philip-William Grassal, Jonas Robl, Florian Kerschbaum
Abstract Releasing data in the form of trained neural networks with differential privacy promises meaningful anonymization. However, there is an inherent privacy-accuracy trade-off in differential privacy which is challenging to assess for non-privacy experts. Furthermore, local and central differential privacy mechanisms are available to either anonymize the training data or the learnt neural network, and the privacy parameter $\epsilon$ cannot be used to compare these two mechanisms. We propose to measure privacy through a black-box membership inference attack and compare the privacy-accuracy trade-off for different local and central differential privacy mechanisms. Furthermore, we need to evaluate whether differential privacy is a useful mechanism in practice since differential privacy will especially be used by data scientists if membership inference risk is lowered more than accuracy. We experiment with several datasets and show that neither local differential privacy nor central differential privacy yields a consistently better privacy-accuracy trade-off in all cases. We also show that the relative privacy-accuracy trade-off, instead of strictly declining linearly over $\epsilon$, is only favorable within a small interval. For this purpose we propose $\varphi$, a ratio expressing the relative privacy-accuracy trade-off.
Tasks Inference Attack
Published 2019-12-24
URL https://arxiv.org/abs/1912.11328v3
PDF https://arxiv.org/pdf/1912.11328v3.pdf
PWC https://paperswithcode.com/paper/assessing-differentially-private-deep
Repo
Framework

Neural-Guided Symbolic Regression with Asymptotic Constraints

Title Neural-Guided Symbolic Regression with Asymptotic Constraints
Authors Li Li, Minjie Fan, Rishabh Singh, Patrick Riley
Abstract Symbolic regression is a type of discrete optimization problem that involves searching expressions that fit given data points. In many cases, other mathematical constraints about the unknown expression not only provide more information beyond just values at some inputs, but also effectively constrain the search space. We identify the asymptotic constraints of leading polynomial powers as the function approaches zero and infinity as useful constraints and create a system to use them for symbolic regression. The first part of the system is a conditional production rule generating neural network which preferentially generates production rules to construct expressions with the desired leading powers, producing novel expressions outside the training domain. The second part, which we call Neural-Guided Monte Carlo Tree Search, uses the network during a search to find an expression that conforms to a set of data points and desired leading powers. Lastly, we provide an extensive experimental validation on thousands of target expressions showing the efficacy of our system compared to exiting methods for finding unknown functions outside of the training set.
Tasks
Published 2019-01-23
URL https://arxiv.org/abs/1901.07714v2
PDF https://arxiv.org/pdf/1901.07714v2.pdf
PWC https://paperswithcode.com/paper/neural-guided-symbolic-regression-with
Repo
Framework

Effects of Differential Privacy and Data Skewness on Membership Inference Vulnerability

Title Effects of Differential Privacy and Data Skewness on Membership Inference Vulnerability
Authors Stacey Truex, Ling Liu, Mehmet Emre Gursoy, Wenqi Wei, Lei Yu
Abstract Membership inference attacks seek to infer the membership of individual training instances of a privately trained model. This paper presents a membership privacy analysis and evaluation system, called MPLens, with three unique contributions. First, through MPLens, we demonstrate how membership inference attack methods can be leveraged in adversarial machine learning. Second, through MPLens, we highlight how the vulnerability of pre-trained models under membership inference attack is not uniform across all classes, particularly when the training data itself is skewed. We show that risk from membership inference attacks is routinely increased when models use skewed training data. Finally, we investigate the effectiveness of differential privacy as a mitigation technique against membership inference attacks. We discuss the trade-offs of implementing such a mitigation strategy with respect to the model complexity, the learning task complexity, the dataset complexity and the privacy parameter settings. Our empirical results reveal that (1) minority groups within skewed datasets display increased risk for membership inference and (2) differential privacy presents many challenging trade-offs as a mitigation technique to membership inference risk.
Tasks Inference Attack
Published 2019-11-21
URL https://arxiv.org/abs/1911.09777v1
PDF https://arxiv.org/pdf/1911.09777v1.pdf
PWC https://paperswithcode.com/paper/effects-of-differential-privacy-and-data
Repo
Framework

Quantifying (Hyper) Parameter Leakage in Machine Learning

Title Quantifying (Hyper) Parameter Leakage in Machine Learning
Authors Vasisht Duddu, D. Vijay Rao
Abstract Machine Learning models, extensively used for various multimedia applications, are offered to users as a blackbox service on the Cloud on a pay-per-query basis. Such blackbox models are commercially valuable to adversaries, making them vulnerable to extraction attacks to reverse engineer the proprietary model thereby violating the model privacy and Intellectual Property. Here, the adversary first extracts the model architecture or hyperparameters through side channel leakage, followed by stealing the functionality of the target model by training the reconstructed architecture on a synthetic dataset. While the attacks proposed in literature are empirical, there is a need for a theoretical framework to measure the information leaked under such extraction attacks. To this extent, in this work, we propose a novel probabilistic framework, Airavata, to estimate the information leakage in such model extraction attacks. This framework captures the fact that extracting the exact target model is difficult due to experimental uncertainty while inferring model hyperparameters and stochastic nature of training to steal the target model functionality. Specifically, we use Bayesian Networks to capture uncertainty in estimating the target model under various extraction attacks based on the subjective notion of probability. We validate the proposed framework under different adversary assumptions commonly adopted in literature to reason about the attack efficacy. This provides a practical tool to infer actionable details about extracting blackbox models and help identify the best attack combination which maximises the knowledge extracted (or information leaked) from the target model.
Tasks Inference Attack
Published 2019-10-31
URL https://arxiv.org/abs/1910.14409v2
PDF https://arxiv.org/pdf/1910.14409v2.pdf
PWC https://paperswithcode.com/paper/quantifying-hyper-parameter-leakage-in
Repo
Framework

Multiple Sample Clustering

Title Multiple Sample Clustering
Authors Xiang Wang, Tie Liu
Abstract The clustering algorithms that view each object data as a single sample drawn from a certain distribution, Gaussian distribution, for example, has been a hot topic for decades. Many clustering algorithms: such as k-means and spectral clustering are proposed based on the single sample assumption. However, in real life, each input object can usually be the multiple samples drawn from a certain hidden distribution. The traditional clustering algorithms cannot handle such a situation. This calls for the multiple sample clustering algorithm. But the traditional multiple sample clustering algorithms can only handle scalar samples or samples from Gaussian distribution. This constrains the application field of multiple sample clustering algorithms. In this paper, we purpose a general framework for multiple sample clustering. Various algorithms can be generated by this framework. We apply two specific cases of this framework: Wasserstein distance version and Bhattacharyya distance version on both synthetic data and stock price data. The simulation results show that the sufficient statistic can greatly improve the clustering accuracy and stability.
Tasks
Published 2019-10-22
URL https://arxiv.org/abs/1910.09731v3
PDF https://arxiv.org/pdf/1910.09731v3.pdf
PWC https://paperswithcode.com/paper/multiple-sample-clustering
Repo
Framework

Robust and Undetectable White-Box Watermarks for Deep Neural Networks

Title Robust and Undetectable White-Box Watermarks for Deep Neural Networks
Authors Tianhao Wang, Florian Kerschbaum
Abstract Watermarking of deep neural networks (DNN) can enable their tracing once released by a data owner. In this paper we generalize white-box watermarking algorithms for DNNs, where the data owner needs white-box access to the model to extract the watermark, and attack and defend them using DNNs. White-box watermarking algorithms have the advantage that they do not impact the accuracy of the watermarked model. We demonstrate a new property inference attack using a DNN that can detect watermarking by any existing, manually designed algorithm regardless of training data set and model architecture. We then use a new training architecture and a further DNN to create a new white-box watermarking algorithm that does not impact accuracy, is undetectable and robust against moderate model transformation attacks.
Tasks Inference Attack
Published 2019-10-31
URL https://arxiv.org/abs/1910.14268v2
PDF https://arxiv.org/pdf/1910.14268v2.pdf
PWC https://paperswithcode.com/paper/robust-and-undetectable-white-box-watermarks
Repo
Framework

Multivariate Spatial Data Visualization: A Survey

Title Multivariate Spatial Data Visualization: A Survey
Authors Xiangyang He, Yubo Tao, Qirui Wang, Hai Lin
Abstract Multivariate spatial data plays an important role in computational science and engineering simulations. The potential features and hidden relationships in multivariate data can assist scientists to gain an in-depth understanding of a scientific process, verify a hypothesis and further discover a new physical or chemical law. In this paper, we present a comprehensive survey of the state-of-the-art techniques for multivariate spatial data visualization. We first introduce the basic concept and characteristics of multivariate spatial data, and describe three main tasks in multivariate data visualization: feature classification, fusion visualization, and correlation analysis. Finally, we prospect potential research topics for multivariate data visualization according to the current research.
Tasks
Published 2019-08-18
URL https://arxiv.org/abs/1908.11344v1
PDF https://arxiv.org/pdf/1908.11344v1.pdf
PWC https://paperswithcode.com/paper/multivariate-spatial-data-visualization-a
Repo
Framework

Oops! Predicting Unintentional Action in Video

Title Oops! Predicting Unintentional Action in Video
Authors Dave Epstein, Boyuan Chen, Carl Vondrick
Abstract From just a short glance at a video, we can often tell whether a person’s action is intentional or not. Can we train a model to recognize this? We introduce a dataset of in-the-wild videos of unintentional action, as well as a suite of tasks for recognizing, localizing, and anticipating its onset. We train a supervised neural network as a baseline and analyze its performance compared to human consistency on the tasks. We also investigate self-supervised representations that leverage natural signals in our dataset, and show the effectiveness of an approach that uses the intrinsic speed of video to perform competitively with highly-supervised pretraining. However, a significant gap between machine and human performance remains. The project website is available at https://oops.cs.columbia.edu
Tasks
Published 2019-11-25
URL https://arxiv.org/abs/1911.11206v1
PDF https://arxiv.org/pdf/1911.11206v1.pdf
PWC https://paperswithcode.com/paper/oops-predicting-unintentional-action-in-video
Repo
Framework

Optimising Optimisers with Push GP

Title Optimising Optimisers with Push GP
Authors Michael Lones
Abstract This work uses Push GP to automatically design both local and population-based optimisers for continuous-valued problems. The optimisers are trained on a single function optimisation landscape, using random transformations to discourage overfitting. They are then tested for generality on larger versions of the same problem, and on other continuous-valued problems. In most cases, the optimisers generalise well to the larger problems. Surprisingly, some of them also generalise very well to previously unseen problems, outperforming existing general purpose optimisers such as CMA-ES. Analysis of the behaviour of the evolved optimisers indicates a range of interesting optimisation strategies that are not found within conventional optimisers, suggesting that this approach could be useful for discovering novel and effective forms of optimisation in an automated manner.
Tasks
Published 2019-10-02
URL https://arxiv.org/abs/1910.00945v1
PDF https://arxiv.org/pdf/1910.00945v1.pdf
PWC https://paperswithcode.com/paper/optimising-optimisers-with-push-gp
Repo
Framework
comments powered by Disqus