Paper Group AWR 138
Differentially Private Releasing via Deep Generative Model (Technical Report). DRCD: a Chinese Machine Reading Comprehension Dataset. Deep Reinforcement Learning for Vision-Based Robotic Grasping: A Simulated Comparative Evaluation of Off-Policy Methods. Density Adaptive Point Set Registration. MIXGAN: Learning Concepts from Different Domains for M …
Differentially Private Releasing via Deep Generative Model (Technical Report)
Title | Differentially Private Releasing via Deep Generative Model (Technical Report) |
Authors | Xinyang Zhang, Shouling Ji, Ting Wang |
Abstract | Privacy-preserving releasing of complex data (e.g., image, text, audio) represents a long-standing challenge for the data mining research community. Due to rich semantics of the data and lack of a priori knowledge about the analysis task, excessive sanitization is often necessary to ensure privacy, leading to significant loss of the data utility. In this paper, we present dp-GAN, a general private releasing framework for semantic-rich data. Instead of sanitizing and then releasing the data, the data curator publishes a deep generative model which is trained using the original data in a differentially private manner; with the generative model, the analyst is able to produce an unlimited amount of synthetic data for arbitrary analysis tasks. In contrast of alternative solutions, dp-GAN highlights a set of key features: (i) it provides theoretical privacy guarantee via enforcing the differential privacy principle; (ii) it retains desirable utility in the released model, enabling a variety of otherwise impossible analyses; and (iii) most importantly, it achieves practical training scalability and stability by employing multi-fold optimization strategies. Through extensive empirical evaluation on benchmark datasets and analyses, we validate the efficacy of dp-GAN. |
Tasks | |
Published | 2018-01-05 |
URL | http://arxiv.org/abs/1801.01594v2 |
http://arxiv.org/pdf/1801.01594v2.pdf | |
PWC | https://paperswithcode.com/paper/differentially-private-releasing-via-deep |
Repo | https://github.com/alexandrehuat/dp-gan |
Framework | tf |
DRCD: a Chinese Machine Reading Comprehension Dataset
Title | DRCD: a Chinese Machine Reading Comprehension Dataset |
Authors | Chih Chieh Shao, Trois Liu, Yuting Lai, Yiying Tseng, Sam Tsai |
Abstract | In this paper, we introduce DRCD (Delta Reading Comprehension Dataset), an open domain traditional Chinese machine reading comprehension (MRC) dataset. This dataset aimed to be a standard Chinese machine reading comprehension dataset, which can be a source dataset in transfer learning. The dataset contains 10,014 paragraphs from 2,108 Wikipedia articles and 30,000+ questions generated by annotators. We build a baseline model that achieves an F1 score of 89.59%. F1 score of Human performance is 93.30%. |
Tasks | Machine Reading Comprehension, Reading Comprehension, Transfer Learning |
Published | 2018-06-04 |
URL | https://arxiv.org/abs/1806.00920v3 |
https://arxiv.org/pdf/1806.00920v3.pdf | |
PWC | https://paperswithcode.com/paper/drcd-a-chinese-machine-reading-comprehension |
Repo | https://github.com/DRCSolutionService/DRCD |
Framework | none |
Deep Reinforcement Learning for Vision-Based Robotic Grasping: A Simulated Comparative Evaluation of Off-Policy Methods
Title | Deep Reinforcement Learning for Vision-Based Robotic Grasping: A Simulated Comparative Evaluation of Off-Policy Methods |
Authors | Deirdre Quillen, Eric Jang, Ofir Nachum, Chelsea Finn, Julian Ibarz, Sergey Levine |
Abstract | In this paper, we explore deep reinforcement learning algorithms for vision-based robotic grasping. Model-free deep reinforcement learning (RL) has been successfully applied to a range of challenging environments, but the proliferation of algorithms makes it difficult to discern which particular approach would be best suited for a rich, diverse task like grasping. To answer this question, we propose a simulated benchmark for robotic grasping that emphasizes off-policy learning and generalization to unseen objects. Off-policy learning enables utilization of grasping data over a wide variety of objects, and diversity is important to enable the method to generalize to new objects that were not seen during training. We evaluate the benchmark tasks against a variety of Q-function estimation methods, a method previously proposed for robotic grasping with deep neural network models, and a novel approach based on a combination of Monte Carlo return estimation and an off-policy correction. Our results indicate that several simple methods provide a surprisingly strong competitor to popular algorithms such as double Q-learning, and our analysis of stability sheds light on the relative tradeoffs between the algorithms. |
Tasks | Q-Learning, Robotic Grasping |
Published | 2018-02-28 |
URL | http://arxiv.org/abs/1802.10264v2 |
http://arxiv.org/pdf/1802.10264v2.pdf | |
PWC | https://paperswithcode.com/paper/deep-reinforcement-learning-for-vision-based |
Repo | https://github.com/mveres01/deepq-grasping |
Framework | pytorch |
Density Adaptive Point Set Registration
Title | Density Adaptive Point Set Registration |
Authors | Felix Järemo Lawin, Martin Danelljan, Fahad Shahbaz Khan, Per-Erik Forssén, Michael Felsberg |
Abstract | Probabilistic methods for point set registration have demonstrated competitive results in recent years. These techniques estimate a probability distribution model of the point clouds. While such a representation has shown promise, it is highly sensitive to variations in the density of 3D points. This fundamental problem is primarily caused by changes in the sensor location across point sets. We revisit the foundations of the probabilistic registration paradigm. Contrary to previous works, we model the underlying structure of the scene as a latent probability distribution, and thereby induce invariance to point set density changes. Both the probabilistic model of the scene and the registration parameters are inferred by minimizing the Kullback-Leibler divergence in an Expectation Maximization based framework. Our density-adaptive registration successfully handles severe density variations commonly encountered in terrestrial Lidar applications. We perform extensive experiments on several challenging real-world Lidar datasets. The results demonstrate that our approach outperforms state-of-the-art probabilistic methods for multi-view registration, without the need of re-sampling. Code is available at https://github.com/felja633/DARE. |
Tasks | |
Published | 2018-04-04 |
URL | http://arxiv.org/abs/1804.01495v2 |
http://arxiv.org/pdf/1804.01495v2.pdf | |
PWC | https://paperswithcode.com/paper/density-adaptive-point-set-registration |
Repo | https://github.com/felja633/DARE |
Framework | none |
MIXGAN: Learning Concepts from Different Domains for Mixture Generation
Title | MIXGAN: Learning Concepts from Different Domains for Mixture Generation |
Authors | Guang-Yuan Hao, Hong-Xing Yu, Wei-Shi Zheng |
Abstract | In this work, we present an interesting attempt on mixture generation: absorbing different image concepts (e.g., content and style) from different domains and thus generating a new domain with learned concepts. In particular, we propose a mixture generative adversarial network (MIXGAN). MIXGAN learns concepts of content and style from two domains respectively, and thus can join them for mixture generation in a new domain, i.e., generating images with content from one domain and style from another. MIXGAN overcomes the limitation of current GAN-based models which either generate new images in the same domain as they observed in training stage, or require off-the-shelf content templates for transferring or translation. Extensive experimental results demonstrate the effectiveness of MIXGAN as compared to related state-of-the-art GAN-based models. |
Tasks | |
Published | 2018-07-04 |
URL | http://arxiv.org/abs/1807.01659v1 |
http://arxiv.org/pdf/1807.01659v1.pdf | |
PWC | https://paperswithcode.com/paper/mixgan-learning-concepts-from-different |
Repo | https://github.com/GuangyuanHao/MIXGAN |
Framework | tf |
Learning Linear Transformations for Fast Arbitrary Style Transfer
Title | Learning Linear Transformations for Fast Arbitrary Style Transfer |
Authors | Xueting Li, Sifei Liu, Jan Kautz, Ming-Hsuan Yang |
Abstract | Given a random pair of images, an arbitrary style transfer method extracts the feel from the reference image to synthesize an output based on the look of the other content image. Recent arbitrary style transfer methods transfer second order statistics from reference image onto content image via a multiplication between content image features and a transformation matrix, which is computed from features with a pre-determined algorithm. These algorithms either require computationally expensive operations, or fail to model the feature covariance and produce artifacts in synthesized images. Generalized from these methods, in this work, we derive the form of transformation matrix theoretically and present an arbitrary style transfer approach that learns the transformation matrix with a feed-forward network. Our algorithm is highly efficient yet allows a flexible combination of multi-level styles while preserving content affinity during style transfer process. We demonstrate the effectiveness of our approach on four tasks: artistic style transfer, video and photo-realistic style transfer as well as domain adaptation, including comparisons with the state-of-the-art methods. |
Tasks | Domain Adaptation, Style Transfer |
Published | 2018-08-14 |
URL | http://arxiv.org/abs/1808.04537v1 |
http://arxiv.org/pdf/1808.04537v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-linear-transformations-for-fast |
Repo | https://github.com/sunshineatnoon/LinearStyleTransfer |
Framework | pytorch |
Learning Stochastic Differential Equations With Gaussian Processes Without Gradient Matching
Title | Learning Stochastic Differential Equations With Gaussian Processes Without Gradient Matching |
Authors | Cagatay Yildiz, Markus Heinonen, Jukka Intosalmi, Henrik Mannerström, Harri Lähdesmäki |
Abstract | We introduce a novel paradigm for learning non-parametric drift and diffusion functions for stochastic differential equation (SDE). The proposed model learns to simulate path distributions that match observations with non-uniform time increments and arbitrary sparseness, which is in contrast with gradient matching that does not optimize simulated responses. We formulate sensitivity equations for learning and demonstrate that our general stochastic distribution optimisation leads to robust and efficient learning of SDE systems. |
Tasks | Gaussian Processes |
Published | 2018-07-16 |
URL | http://arxiv.org/abs/1807.05748v2 |
http://arxiv.org/pdf/1807.05748v2.pdf | |
PWC | https://paperswithcode.com/paper/learning-stochastic-differential-equations |
Repo | https://github.com/cagatayyildiz/npde |
Framework | tf |
Conversations Gone Awry: Detecting Early Signs of Conversational Failure
Title | Conversations Gone Awry: Detecting Early Signs of Conversational Failure |
Authors | Justine Zhang, Jonathan P. Chang, Cristian Danescu-Niculescu-Mizil, Lucas Dixon, Yiqing Hua, Nithum Thain, Dario Taraborelli |
Abstract | One of the main challenges online social systems face is the prevalence of antisocial behavior, such as harassment and personal attacks. In this work, we introduce the task of predicting from the very start of a conversation whether it will get out of hand. As opposed to detecting undesirable behavior after the fact, this task aims to enable early, actionable prediction at a time when the conversation might still be salvaged. To this end, we develop a framework for capturing pragmatic devices—such as politeness strategies and rhetorical prompts—used to start a conversation, and analyze their relation to its future trajectory. Applying this framework in a controlled setting, we demonstrate the feasibility of detecting early warning signs of antisocial behavior in online discussions. |
Tasks | |
Published | 2018-05-14 |
URL | http://arxiv.org/abs/1805.05345v1 |
http://arxiv.org/pdf/1805.05345v1.pdf | |
PWC | https://paperswithcode.com/paper/conversations-gone-awry-detecting-early-signs |
Repo | https://github.com/canonical-debate-lab/paper |
Framework | none |
Learning to Perform Local Rewriting for Combinatorial Optimization
Title | Learning to Perform Local Rewriting for Combinatorial Optimization |
Authors | Xinyun Chen, Yuandong Tian |
Abstract | Search-based methods for hard combinatorial optimization are often guided by heuristics. Tuning heuristics in various conditions and situations is often time-consuming. In this paper, we propose NeuRewriter that learns a policy to pick heuristics and rewrite the local components of the current solution to iteratively improve it until convergence. The policy factorizes into a region-picking and a rule-picking component, each parameterized by a neural network trained with actor-critic methods in reinforcement learning. NeuRewriter captures the general structure of combinatorial problems and shows strong performance in three versatile tasks: expression simplification, online job scheduling and vehicle routing problems. NeuRewriter outperforms the expression simplification component in Z3; outperforms DeepRM and Google OR-tools in online job scheduling; and outperforms recent neural baselines and Google OR-tools in vehicle routing problems. |
Tasks | Combinatorial Optimization |
Published | 2018-09-30 |
URL | https://arxiv.org/abs/1810.00337v5 |
https://arxiv.org/pdf/1810.00337v5.pdf | |
PWC | https://paperswithcode.com/paper/automatic-local-rewriting-for-combinatorial |
Repo | https://github.com/facebookresearch/neural-rewriter |
Framework | pytorch |
Invertible Residual Networks
Title | Invertible Residual Networks |
Authors | Jens Behrmann, Will Grathwohl, Ricky T. Q. Chen, David Duvenaud, Jörn-Henrik Jacobsen |
Abstract | We show that standard ResNet architectures can be made invertible, allowing the same model to be used for classification, density estimation, and generation. Typically, enforcing invertibility requires partitioning dimensions or restricting network architectures. In contrast, our approach only requires adding a simple normalization step during training, already available in standard frameworks. Invertible ResNets define a generative model which can be trained by maximum likelihood on unlabeled data. To compute likelihoods, we introduce a tractable approximation to the Jacobian log-determinant of a residual block. Our empirical evaluation shows that invertible ResNets perform competitively with both state-of-the-art image classifiers and flow-based generative models, something that has not been previously achieved with a single architecture. |
Tasks | Density Estimation, Image Generation |
Published | 2018-11-02 |
URL | https://arxiv.org/abs/1811.00995v3 |
https://arxiv.org/pdf/1811.00995v3.pdf | |
PWC | https://paperswithcode.com/paper/invertible-residual-networks |
Repo | https://github.com/rtqichen/residual-flows |
Framework | pytorch |
Character-based Neural Networks for Sentence Pair Modeling
Title | Character-based Neural Networks for Sentence Pair Modeling |
Authors | Wuwei Lan, Wei Xu |
Abstract | Sentence pair modeling is critical for many NLP tasks, such as paraphrase identification, semantic textual similarity, and natural language inference. Most state-of-the-art neural models for these tasks rely on pretrained word embedding and compose sentence-level semantics in varied ways; however, few works have attempted to verify whether we really need pretrained embeddings in these tasks. In this paper, we study how effective subword-level (character and character n-gram) representations are in sentence pair modeling. Though it is well-known that subword models are effective in tasks with single sentence input, including language modeling and machine translation, they have not been systematically studied in sentence pair modeling tasks where the semantic and string similarities between texts matter. Our experiments show that subword models without any pretrained word embedding can achieve new state-of-the-art results on two social media datasets and competitive results on news data for paraphrase identification. |
Tasks | Language Modelling, Natural Language Inference, Paraphrase Identification, Semantic Textual Similarity, Sentence Pair Modeling |
Published | 2018-05-21 |
URL | http://arxiv.org/abs/1805.08297v1 |
http://arxiv.org/pdf/1805.08297v1.pdf | |
PWC | https://paperswithcode.com/paper/character-based-neural-networks-for-sentence |
Repo | https://github.com/lanwuwei/SPM_toolkit |
Framework | pytorch |
Void Filling of Digital Elevation Models with Deep Generative Models
Title | Void Filling of Digital Elevation Models with Deep Generative Models |
Authors | Konstantinos Gavriil, Georg Muntingh, Oliver J. D. Barrowclough |
Abstract | In recent years, advances in machine learning algorithms, cheap computational resources, and the availability of big data have spurred the deep learning revolution in various application domains. In particular, supervised learning techniques in image analysis have led to superhuman performance in various tasks, such as classification, localization, and segmentation, while unsupervised learning techniques based on increasingly advanced generative models have been applied to generate high-resolution synthetic images indistinguishable from real images. In this paper we consider a state-of-the-art machine learning model for image inpainting, namely a Wasserstein Generative Adversarial Network based on a fully convolutional architecture with a contextual attention mechanism. We show that this model can successfully be transferred to the setting of digital elevation models (DEMs) for the purpose of generating semantically plausible data for filling voids. Training, testing and experimentation is done on GeoTIFF data from various regions in Norway, made openly available by the Norwegian Mapping Authority. |
Tasks | Image Inpainting |
Published | 2018-11-30 |
URL | http://arxiv.org/abs/1811.12693v2 |
http://arxiv.org/pdf/1811.12693v2.pdf | |
PWC | https://paperswithcode.com/paper/void-filling-of-digital-elevation-models-with |
Repo | https://github.com/konstantg/dem-fill |
Framework | tf |
Faster Matrix Completion Using Randomized SVD
Title | Faster Matrix Completion Using Randomized SVD |
Authors | Xu Feng, Wenjian Yu, Yaohang Li |
Abstract | Matrix completion is a widely used technique for image inpainting and personalized recommender system, etc. In this work, we focus on accelerating the matrix completion using faster randomized singular value decomposition (rSVD). Firstly, two fast randomized algorithms (rSVD-PI and rSVD- BKI) are proposed for handling sparse matrix. They make use of an eigSVD procedure and several accelerating skills. Then, with the rSVD-BKI algorithm and a new subspace recycling technique, we accelerate the singular value thresholding (SVT) method in [1] to realize faster matrix completion. Experiments show that the proposed rSVD algorithms can be 6X faster than the basic rSVD algorithm [2] while keeping same accuracy. For image inpainting and movie-rating estimation problems, the proposed accelerated SVT algorithm consumes 15X and 8X less CPU time than the methods using svds and lansvd respectively, without loss of accuracy. |
Tasks | Image Inpainting, Matrix Completion, Recommendation Systems |
Published | 2018-10-16 |
URL | http://arxiv.org/abs/1810.06860v1 |
http://arxiv.org/pdf/1810.06860v1.pdf | |
PWC | https://paperswithcode.com/paper/faster-matrix-completion-using-randomized-svd |
Repo | https://github.com/XuFengthucs/fSVT |
Framework | none |
Speech2Vec: A Sequence-to-Sequence Framework for Learning Word Embeddings from Speech
Title | Speech2Vec: A Sequence-to-Sequence Framework for Learning Word Embeddings from Speech |
Authors | Yu-An Chung, James Glass |
Abstract | In this paper, we propose a novel deep neural network architecture, Speech2Vec, for learning fixed-length vector representations of audio segments excised from a speech corpus, where the vectors contain semantic information pertaining to the underlying spoken words, and are close to other vectors in the embedding space if their corresponding underlying spoken words are semantically similar. The proposed model can be viewed as a speech version of Word2Vec. Its design is based on a RNN Encoder-Decoder framework, and borrows the methodology of skipgrams or continuous bag-of-words for training. Learning word embeddings directly from speech enables Speech2Vec to make use of the semantic information carried by speech that does not exist in plain text. The learned word embeddings are evaluated and analyzed on 13 widely used word similarity benchmarks, and outperform word embeddings learned by Word2Vec from the transcriptions. |
Tasks | Learning Word Embeddings, Word Embeddings |
Published | 2018-03-23 |
URL | http://arxiv.org/abs/1803.08976v2 |
http://arxiv.org/pdf/1803.08976v2.pdf | |
PWC | https://paperswithcode.com/paper/speech2vec-a-sequence-to-sequence-framework |
Repo | https://github.com/iamyuanchung/speech2vec-pretrained-vectors |
Framework | none |
A Hybrid Frequency-domain/Image-domain Deep Network for Magnetic Resonance Image Reconstruction
Title | A Hybrid Frequency-domain/Image-domain Deep Network for Magnetic Resonance Image Reconstruction |
Authors | Roberto Souza, Richard Frayne |
Abstract | Decreasing magnetic resonance (MR) image acquisition times can potentially reduce procedural cost and make MR examinations more accessible. Compressed sensing (CS)-based image reconstruction methods, for example, decrease MR acquisition time by reconstructing high-quality images from data that were originally sampled at rates inferior to the Nyquist-Shannon sampling theorem. In this work we propose a hybrid architecture that works both in the k-space (or frequency-domain) and the image (or spatial) domains. Our network is composed of a complex-valued residual U-net in the k-space domain, an inverse Fast Fourier Transform (iFFT) operation, and a real-valued U-net in the image domain. Our experiments demonstrated, using MR raw k-space data, that the proposed hybrid approach can potentially improve CS reconstruction compared to deep-learning networks that operate only in the image domain. In this study we compare our method with four previously published deep neural networks and examine their ability to reconstruct images that are subsequently used to generate regional volume estimates. We evaluated undersampling ratios of 75% and 80%. Our technique was ranked second in the quantitative analysis, but qualitative analysis indicated that our reconstruction performed the best in hard to reconstruct regions, such as the cerebellum. All images reconstructed with our method were successfully post-processed, and showed good volumetry agreement compared with the fully sampled reconstruction measures. |
Tasks | Image Reconstruction |
Published | 2018-10-30 |
URL | http://arxiv.org/abs/1810.12473v1 |
http://arxiv.org/pdf/1810.12473v1.pdf | |
PWC | https://paperswithcode.com/paper/a-hybrid-frequency-domainimage-domain-deep |
Repo | https://github.com/rmsouza01/Hybrid-CS-Model-MRI |
Framework | tf |