Paper Group AWR 28
Neural Sketch Learning for Conditional Program Generation. ZOO: Zeroth Order Optimization based Black-box Attacks to Deep Neural Networks without Training Substitute Models. W-Net: A Deep Model for Fully Unsupervised Image Segmentation. Efficient Yet Deep Convolutional Neural Networks for Semantic Segmentation. Wembedder: Wikidata entity embedding …
Neural Sketch Learning for Conditional Program Generation
Title | Neural Sketch Learning for Conditional Program Generation |
Authors | Vijayaraghavan Murali, Letao Qi, Swarat Chaudhuri, Chris Jermaine |
Abstract | We study the problem of generating source code in a strongly typed, Java-like programming language, given a label (for example a set of API calls or types) carrying a small amount of information about the code that is desired. The generated programs are expected to respect a “realistic” relationship between programs and labels, as exemplified by a corpus of labeled programs available during training. Two challenges in such conditional program generation are that the generated programs must satisfy a rich set of syntactic and semantic constraints, and that source code contains many low-level features that impede learning. We address these problems by training a neural generator not on code but on program sketches, or models of program syntax that abstract out names and operations that do not generalize across programs. During generation, we infer a posterior distribution over sketches, then concretize samples from this distribution into type-safe programs using combinatorial techniques. We implement our ideas in a system for generating API-heavy Java code, and show that it can often predict the entire body of a method given just a few API calls or data types that appear in the method. |
Tasks | |
Published | 2017-03-16 |
URL | http://arxiv.org/abs/1703.05698v5 |
http://arxiv.org/pdf/1703.05698v5.pdf | |
PWC | https://paperswithcode.com/paper/neural-sketch-learning-for-conditional |
Repo | https://github.com/capergroup/bayou |
Framework | tf |
ZOO: Zeroth Order Optimization based Black-box Attacks to Deep Neural Networks without Training Substitute Models
Title | ZOO: Zeroth Order Optimization based Black-box Attacks to Deep Neural Networks without Training Substitute Models |
Authors | Pin-Yu Chen, Huan Zhang, Yash Sharma, Jinfeng Yi, Cho-Jui Hsieh |
Abstract | Deep neural networks (DNNs) are one of the most prominent technologies of our time, as they achieve state-of-the-art performance in many machine learning tasks, including but not limited to image classification, text mining, and speech processing. However, recent research on DNNs has indicated ever-increasing concern on the robustness to adversarial examples, especially for security-critical tasks such as traffic sign identification for autonomous driving. Studies have unveiled the vulnerability of a well-trained DNN by demonstrating the ability of generating barely noticeable (to both human and machines) adversarial images that lead to misclassification. Furthermore, researchers have shown that these adversarial images are highly transferable by simply training and attacking a substitute model built upon the target model, known as a black-box attack to DNNs. Similar to the setting of training substitute models, in this paper we propose an effective black-box attack that also only has access to the input (images) and the output (confidence scores) of a targeted DNN. However, different from leveraging attack transferability from substitute models, we propose zeroth order optimization (ZOO) based attacks to directly estimate the gradients of the targeted DNN for generating adversarial examples. We use zeroth order stochastic coordinate descent along with dimension reduction, hierarchical attack and importance sampling techniques to efficiently attack black-box models. By exploiting zeroth order optimization, improved attacks to the targeted DNN can be accomplished, sparing the need for training substitute models and avoiding the loss in attack transferability. Experimental results on MNIST, CIFAR10 and ImageNet show that the proposed ZOO attack is as effective as the state-of-the-art white-box attack and significantly outperforms existing black-box attacks via substitute models. |
Tasks | Adversarial Attack, Adversarial Defense, Autonomous Driving, Dimensionality Reduction, Image Classification |
Published | 2017-08-14 |
URL | http://arxiv.org/abs/1708.03999v2 |
http://arxiv.org/pdf/1708.03999v2.pdf | |
PWC | https://paperswithcode.com/paper/zoo-zeroth-order-optimization-based-black-box |
Repo | https://github.com/huanzhang12/ZOO-Attack |
Framework | tf |
W-Net: A Deep Model for Fully Unsupervised Image Segmentation
Title | W-Net: A Deep Model for Fully Unsupervised Image Segmentation |
Authors | Xide Xia, Brian Kulis |
Abstract | While significant attention has been recently focused on designing supervised deep semantic segmentation algorithms for vision tasks, there are many domains in which sufficient supervised pixel-level labels are difficult to obtain. In this paper, we revisit the problem of purely unsupervised image segmentation and propose a novel deep architecture for this problem. We borrow recent ideas from supervised semantic segmentation methods, in particular by concatenating two fully convolutional networks together into an autoencoder–one for encoding and one for decoding. The encoding layer produces a k-way pixelwise prediction, and both the reconstruction error of the autoencoder as well as the normalized cut produced by the encoder are jointly minimized during training. When combined with suitable postprocessing involving conditional random field smoothing and hierarchical segmentation, our resulting algorithm achieves impressive results on the benchmark Berkeley Segmentation Data Set, outperforming a number of competing methods. |
Tasks | Semantic Segmentation |
Published | 2017-11-22 |
URL | http://arxiv.org/abs/1711.08506v1 |
http://arxiv.org/pdf/1711.08506v1.pdf | |
PWC | https://paperswithcode.com/paper/w-net-a-deep-model-for-fully-unsupervised |
Repo | https://github.com/wau/Unsupervised-SIS |
Framework | none |
Efficient Yet Deep Convolutional Neural Networks for Semantic Segmentation
Title | Efficient Yet Deep Convolutional Neural Networks for Semantic Segmentation |
Authors | Sharif Amit Kamran, Ali Shihab Sabbir |
Abstract | Semantic Segmentation using deep convolutional neural network pose more complex challenge for any GPU intensive task. As it has to compute million of parameters, it results to huge memory consumption. Moreover, extracting finer features and conducting supervised training tends to increase the complexity. With the introduction of Fully Convolutional Neural Network, which uses finer strides and utilizes deconvolutional layers for upsampling, it has been a go to for any image segmentation task. In this paper, we propose two segmentation architecture which not only needs one-third the parameters to compute but also gives better accuracy than the similar architectures. The model weights were transferred from the popular neural net like VGG19 and VGG16 which were trained on Imagenet classification data-set. Then we transform all the fully connected layers to convolutional layers and use dilated convolution for decreasing the parameters. Lastly, we add finer strides and attach four skip architectures which are element-wise summed with the deconvolutional layers in steps. We train and test on different sparse and fine data-sets like Pascal VOC2012, Pascal-Context and NYUDv2 and show how better our model performs in this tasks. On the other hand our model has a faster inference time and consumes less memory for training and testing on NVIDIA Pascal GPUs, making it more efficient and less memory consuming architecture for pixel-wise segmentation. |
Tasks | Scene Segmentation, Semantic Segmentation |
Published | 2017-07-26 |
URL | http://arxiv.org/abs/1707.08254v3 |
http://arxiv.org/pdf/1707.08254v3.pdf | |
PWC | https://paperswithcode.com/paper/efficient-yet-deep-convolutional-neural |
Repo | https://github.com/SharifAmit/DilatedFCNSegmentation |
Framework | caffe2 |
Wembedder: Wikidata entity embedding web service
Title | Wembedder: Wikidata entity embedding web service |
Authors | Finn Årup Nielsen |
Abstract | I present a web service for querying an embedding of entities in the Wikidata knowledge graph. The embedding is trained on the Wikidata dump using Gensim’s Word2Vec implementation and a simple graph walk. A REST API is implemented. Together with the Wikidata API the web service exposes a multilingual resource for over 600’000 Wikidata items and properties. |
Tasks | |
Published | 2017-10-11 |
URL | http://arxiv.org/abs/1710.04099v1 |
http://arxiv.org/pdf/1710.04099v1.pdf | |
PWC | https://paperswithcode.com/paper/wembedder-wikidata-entity-embedding-web |
Repo | https://github.com/fnielsen/wembedder |
Framework | none |
Channel Pruning for Accelerating Very Deep Neural Networks
Title | Channel Pruning for Accelerating Very Deep Neural Networks |
Authors | Yihui He, Xiangyu Zhang, Jian Sun |
Abstract | In this paper, we introduce a new channel pruning method to accelerate very deep convolutional neural networks.Given a trained CNN model, we propose an iterative two-step algorithm to effectively prune each layer, by a LASSO regression based channel selection and least square reconstruction. We further generalize this algorithm to multi-layer and multi-branch cases. Our method reduces the accumulated error and enhance the compatibility with various architectures. Our pruned VGG-16 achieves the state-of-the-art results by 5x speed-up along with only 0.3% increase of error. More importantly, our method is able to accelerate modern networks like ResNet, Xception and suffers only 1.4%, 1.0% accuracy loss under 2x speed-up respectively, which is significant. Code has been made publicly available. |
Tasks | |
Published | 2017-07-19 |
URL | http://arxiv.org/abs/1707.06168v2 |
http://arxiv.org/pdf/1707.06168v2.pdf | |
PWC | https://paperswithcode.com/paper/channel-pruning-for-accelerating-very-deep |
Repo | https://github.com/yihui-he/channel-pruning |
Framework | none |
Deep Photo Style Transfer
Title | Deep Photo Style Transfer |
Authors | Fujun Luan, Sylvain Paris, Eli Shechtman, Kavita Bala |
Abstract | This paper introduces a deep-learning approach to photographic style transfer that handles a large variety of image content while faithfully transferring the reference style. Our approach builds upon the recent work on painterly transfer that separates style from the content of an image by considering different layers of a neural network. However, as is, this approach is not suitable for photorealistic style transfer. Even when both the input and reference images are photographs, the output still exhibits distortions reminiscent of a painting. Our contribution is to constrain the transformation from the input to the output to be locally affine in colorspace, and to express this constraint as a custom fully differentiable energy term. We show that this approach successfully suppresses distortion and yields satisfying photorealistic style transfers in a broad variety of scenarios, including transfer of the time of day, weather, season, and artistic edits. |
Tasks | Style Transfer |
Published | 2017-03-22 |
URL | http://arxiv.org/abs/1703.07511v3 |
http://arxiv.org/pdf/1703.07511v3.pdf | |
PWC | https://paperswithcode.com/paper/deep-photo-style-transfer |
Repo | https://github.com/brstream/links |
Framework | none |
Deep Unsupervised Clustering Using Mixture of Autoencoders
Title | Deep Unsupervised Clustering Using Mixture of Autoencoders |
Authors | Dejiao Zhang, Yifan Sun, Brian Eriksson, Laura Balzano |
Abstract | Unsupervised clustering is one of the most fundamental challenges in machine learning. A popular hypothesis is that data are generated from a union of low-dimensional nonlinear manifolds; thus an approach to clustering is identifying and separating these manifolds. In this paper, we present a novel approach to solve this problem by using a mixture of autoencoders. Our model consists of two parts: 1) a collection of autoencoders where each autoencoder learns the underlying manifold of a group of similar objects, and 2) a mixture assignment neural network, which takes the concatenated latent vectors from the autoencoders as input and infers the distribution over clusters. By jointly optimizing the two parts, we simultaneously assign data to clusters and learn the underlying manifolds of each cluster. |
Tasks | |
Published | 2017-12-21 |
URL | http://arxiv.org/abs/1712.07788v2 |
http://arxiv.org/pdf/1712.07788v2.pdf | |
PWC | https://paperswithcode.com/paper/deep-unsupervised-clustering-using-mixture-of |
Repo | https://github.com/icannos/mixture-autoencoder |
Framework | tf |
A Read-Write Memory Network for Movie Story Understanding
Title | A Read-Write Memory Network for Movie Story Understanding |
Authors | Seil Na, Sangho Lee, Jisung Kim, Gunhee Kim |
Abstract | We propose a novel memory network model named Read-Write Memory Network (RWMN) to perform question and answering tasks for large-scale, multimodal movie story understanding. The key focus of our RWMN model is to design the read network and the write network that consist of multiple convolutional layers, which enable memory read and write operations to have high capacity and flexibility. While existing memory-augmented network models treat each memory slot as an independent block, our use of multi-layered CNNs allows the model to read and write sequential memory cells as chunks, which is more reasonable to represent a sequential story because adjacent memory blocks often have strong correlations. For evaluation, we apply our model to all the six tasks of the MovieQA benchmark, and achieve the best accuracies on several tasks, especially on the visual QA task. Our model shows a potential to better understand not only the content in the story, but also more abstract information, such as relationships between characters and the reasons for their actions. |
Tasks | Video Story QA |
Published | 2017-09-27 |
URL | http://arxiv.org/abs/1709.09345v4 |
http://arxiv.org/pdf/1709.09345v4.pdf | |
PWC | https://paperswithcode.com/paper/a-read-write-memory-network-for-movie-story |
Repo | https://github.com/seilna/RWMN |
Framework | tf |
Neural Style Transfer: A Review
Title | Neural Style Transfer: A Review |
Authors | Yongcheng Jing, Yezhou Yang, Zunlei Feng, Jingwen Ye, Yizhou Yu, Mingli Song |
Abstract | The seminal work of Gatys et al. demonstrated the power of Convolutional Neural Networks (CNNs) in creating artistic imagery by separating and recombining image content and style. This process of using CNNs to render a content image in different styles is referred to as Neural Style Transfer (NST). Since then, NST has become a trending topic both in academic literature and industrial applications. It is receiving increasing attention and a variety of approaches are proposed to either improve or extend the original NST algorithm. In this paper, we aim to provide a comprehensive overview of the current progress towards NST. We first propose a taxonomy of current algorithms in the field of NST. Then, we present several evaluation methods and compare different NST algorithms both qualitatively and quantitatively. The review concludes with a discussion of various applications of NST and open problems for future research. A list of papers discussed in this review, corresponding codes, pre-trained models and more comparison results are publicly available at https://github.com/ycjing/Neural-Style-Transfer-Papers. |
Tasks | Style Transfer |
Published | 2017-05-11 |
URL | http://arxiv.org/abs/1705.04058v7 |
http://arxiv.org/pdf/1705.04058v7.pdf | |
PWC | https://paperswithcode.com/paper/neural-style-transfer-a-review |
Repo | https://github.com/andy-yangz/writing_style_transfer |
Framework | none |
Deep Neural Networks as Gaussian Processes
Title | Deep Neural Networks as Gaussian Processes |
Authors | Jaehoon Lee, Yasaman Bahri, Roman Novak, Samuel S. Schoenholz, Jeffrey Pennington, Jascha Sohl-Dickstein |
Abstract | It has long been known that a single-layer fully-connected neural network with an i.i.d. prior over its parameters is equivalent to a Gaussian process (GP), in the limit of infinite network width. This correspondence enables exact Bayesian inference for infinite width neural networks on regression tasks by means of evaluating the corresponding GP. Recently, kernel functions which mimic multi-layer random neural networks have been developed, but only outside of a Bayesian framework. As such, previous work has not identified that these kernels can be used as covariance functions for GPs and allow fully Bayesian prediction with a deep neural network. In this work, we derive the exact equivalence between infinitely wide deep networks and GPs. We further develop a computationally efficient pipeline to compute the covariance function for these GPs. We then use the resulting GPs to perform Bayesian inference for wide deep neural networks on MNIST and CIFAR-10. We observe that trained neural network accuracy approaches that of the corresponding GP with increasing layer width, and that the GP uncertainty is strongly correlated with trained network prediction error. We further find that test performance increases as finite-width trained networks are made wider and more similar to a GP, and thus that GP predictions typically outperform those of finite-width networks. Finally we connect the performance of these GPs to the recent theory of signal propagation in random neural networks. |
Tasks | Bayesian Inference, Gaussian Processes |
Published | 2017-11-01 |
URL | http://arxiv.org/abs/1711.00165v3 |
http://arxiv.org/pdf/1711.00165v3.pdf | |
PWC | https://paperswithcode.com/paper/deep-neural-networks-as-gaussian-processes |
Repo | https://github.com/yumaloop/dnn_hsic |
Framework | tf |
Escape from Cells: Deep Kd-Networks for the Recognition of 3D Point Cloud Models
Title | Escape from Cells: Deep Kd-Networks for the Recognition of 3D Point Cloud Models |
Authors | Roman Klokov, Victor Lempitsky |
Abstract | We present a new deep learning architecture (called Kd-network) that is designed for 3D model recognition tasks and works with unstructured point clouds. The new architecture performs multiplicative transformations and share parameters of these transformations according to the subdivisions of the point clouds imposed onto them by Kd-trees. Unlike the currently dominant convolutional architectures that usually require rasterization on uniform two-dimensional or three-dimensional grids, Kd-networks do not rely on such grids in any way and therefore avoid poor scaling behaviour. In a series of experiments with popular shape recognition benchmarks, Kd-networks demonstrate competitive performance in a number of shape recognition tasks such as shape classification, shape retrieval and shape part segmentation. |
Tasks | 3D Part Segmentation |
Published | 2017-04-04 |
URL | http://arxiv.org/abs/1704.01222v2 |
http://arxiv.org/pdf/1704.01222v2.pdf | |
PWC | https://paperswithcode.com/paper/escape-from-cells-deep-kd-networks-for-the |
Repo | https://github.com/fxia22/kdnet.pytorch |
Framework | pytorch |
Like What You Like: Knowledge Distill via Neuron Selectivity Transfer
Title | Like What You Like: Knowledge Distill via Neuron Selectivity Transfer |
Authors | Zehao Huang, Naiyan Wang |
Abstract | Despite deep neural networks have demonstrated extraordinary power in various applications, their superior performances are at expense of high storage and computational costs. Consequently, the acceleration and compression of neural networks have attracted much attention recently. Knowledge Transfer (KT), which aims at training a smaller student network by transferring knowledge from a larger teacher model, is one of the popular solutions. In this paper, we propose a novel knowledge transfer method by treating it as a distribution matching problem. Particularly, we match the distributions of neuron selectivity patterns between teacher and student networks. To achieve this goal, we devise a new KT loss function by minimizing the Maximum Mean Discrepancy (MMD) metric between these distributions. Combined with the original loss function, our method can significantly improve the performance of student networks. We validate the effectiveness of our method across several datasets, and further combine it with other KT methods to explore the best possible results. Last but not least, we fine-tune the model to other tasks such as object detection. The results are also encouraging, which confirm the transferability of the learned features. |
Tasks | Object Detection, Transfer Learning |
Published | 2017-07-05 |
URL | http://arxiv.org/abs/1707.01219v2 |
http://arxiv.org/pdf/1707.01219v2.pdf | |
PWC | https://paperswithcode.com/paper/like-what-you-like-knowledge-distill-via |
Repo | https://github.com/TuSimple/neuron-selectivity-transfer |
Framework | tf |
Learning model-based planning from scratch
Title | Learning model-based planning from scratch |
Authors | Razvan Pascanu, Yujia Li, Oriol Vinyals, Nicolas Heess, Lars Buesing, Sebastien Racanière, David Reichert, Théophane Weber, Daan Wierstra, Peter Battaglia |
Abstract | Conventional wisdom holds that model-based planning is a powerful approach to sequential decision-making. It is often very challenging in practice, however, because while a model can be used to evaluate a plan, it does not prescribe how to construct a plan. Here we introduce the “Imagination-based Planner”, the first model-based, sequential decision-making agent that can learn to construct, evaluate, and execute plans. Before any action, it can perform a variable number of imagination steps, which involve proposing an imagined action and evaluating it with its model-based imagination. All imagined actions and outcomes are aggregated, iteratively, into a “plan context” which conditions future real and imagined actions. The agent can even decide how to imagine: testing out alternative imagined actions, chaining sequences of actions together, or building a more complex “imagination tree” by navigating flexibly among the previously imagined states using a learned policy. And our agent can learn to plan economically, jointly optimizing for external rewards and computational costs associated with using its imagination. We show that our architecture can learn to solve a challenging continuous control problem, and also learn elaborate planning strategies in a discrete maze-solving task. Our work opens a new direction toward learning the components of a model-based planning system and how to use them. |
Tasks | Continuous Control, Decision Making |
Published | 2017-07-19 |
URL | http://arxiv.org/abs/1707.06170v1 |
http://arxiv.org/pdf/1707.06170v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-model-based-planning-from-scratch |
Repo | https://github.com/Driesssens/ibp-pytorch |
Framework | pytorch |
Towards Generalization and Simplicity in Continuous Control
Title | Towards Generalization and Simplicity in Continuous Control |
Authors | Aravind Rajeswaran, Kendall Lowrey, Emanuel Todorov, Sham Kakade |
Abstract | This work shows that policies with simple linear and RBF parameterizations can be trained to solve a variety of continuous control tasks, including the OpenAI gym benchmarks. The performance of these trained policies are competitive with state of the art results, obtained with more elaborate parameterizations such as fully connected neural networks. Furthermore, existing training and testing scenarios are shown to be very limited and prone to over-fitting, thus giving rise to only trajectory-centric policies. Training with a diverse initial state distribution is shown to produce more global policies with better generalization. This allows for interactive control scenarios where the system recovers from large on-line perturbations; as shown in the supplementary video. |
Tasks | Continuous Control |
Published | 2017-03-08 |
URL | http://arxiv.org/abs/1703.02660v2 |
http://arxiv.org/pdf/1703.02660v2.pdf | |
PWC | https://paperswithcode.com/paper/towards-generalization-and-simplicity-in |
Repo | https://github.com/cvoelcker/reinforcement-learning-material-ws-2018 |
Framework | none |