Paper Group AWR 76
Deep Quality-Value (DQV) Learning. AnatomyNet: Deep Learning for Fast and Fully Automated Whole-volume Segmentation of Head and Neck Anatomy. Regularized Wasserstein Means for Aligning Distributional Data. Metric on Nonlinear Dynamical Systems with Perron-Frobenius Operators. Recurrent Neural Networks for Fuzz Testing Web Browsers. Parsimonious Bay …
Deep Quality-Value (DQV) Learning
Title | Deep Quality-Value (DQV) Learning |
Authors | Matthia Sabatelli, Gilles Louppe, Pierre Geurts, Marco A. Wiering |
Abstract | We introduce a novel Deep Reinforcement Learning (DRL) algorithm called Deep Quality-Value (DQV) Learning. DQV uses temporal-difference learning to train a Value neural network and uses this network for training a second Quality-value network that learns to estimate state-action values. We first test DQV’s update rules with Multilayer Perceptrons as function approximators on two classic RL problems, and then extend DQV with the use of Deep Convolutional Neural Networks, Experience Replay' and Target Neural Networks’ for tackling four games of the Atari Arcade Learning environment. Our results show that DQV learns significantly faster and better than Deep Q-Learning and Double Deep Q-Learning, suggesting that our algorithm can potentially be a better performing synchronous temporal difference algorithm than what is currently present in DRL. |
Tasks | Atari Games, Q-Learning |
Published | 2018-09-30 |
URL | http://arxiv.org/abs/1810.00368v2 |
http://arxiv.org/pdf/1810.00368v2.pdf | |
PWC | https://paperswithcode.com/paper/deep-quality-value-dqv-learning |
Repo | https://github.com/paintception/Deep-Quality-Value-DQV-Learning- |
Framework | tf |
AnatomyNet: Deep Learning for Fast and Fully Automated Whole-volume Segmentation of Head and Neck Anatomy
Title | AnatomyNet: Deep Learning for Fast and Fully Automated Whole-volume Segmentation of Head and Neck Anatomy |
Authors | Wentao Zhu, Yufang Huang, Liang Zeng, Xuming Chen, Yong Liu, Zhen Qian, Nan Du, Wei Fan, Xiaohui Xie |
Abstract | Methods: Our deep learning model, called AnatomyNet, segments OARs from head and neck CT images in an end-to-end fashion, receiving whole-volume HaN CT images as input and generating masks of all OARs of interest in one shot. AnatomyNet is built upon the popular 3D U-net architecture, but extends it in three important ways: 1) a new encoding scheme to allow auto-segmentation on whole-volume CT images instead of local patches or subsets of slices, 2) incorporating 3D squeeze-and-excitation residual blocks in encoding layers for better feature representation, and 3) a new loss function combining Dice scores and focal loss to facilitate the training of the neural model. These features are designed to address two main challenges in deep-learning-based HaN segmentation: a) segmenting small anatomies (i.e., optic chiasm and optic nerves) occupying only a few slices, and b) training with inconsistent data annotations with missing ground truth for some anatomical structures. Results: We collected 261 HaN CT images to train AnatomyNet, and used MICCAI Head and Neck Auto Segmentation Challenge 2015 as a benchmark dataset to evaluate the performance of AnatomyNet. The objective is to segment nine anatomies: brain stem, chiasm, mandible, optic nerve left, optic nerve right, parotid gland left, parotid gland right, submandibular gland left, and submandibular gland right. Compared to previous state-of-the-art results from the MICCAI 2015 competition, AnatomyNet increases Dice similarity coefficient by 3.3% on average. AnatomyNet takes about 0.12 seconds to fully segment a head and neck CT image of dimension 178 x 302 x 225, significantly faster than previous methods. In addition, the model is able to process whole-volume CT images and delineate all OARs in one pass, requiring little pre- or post-processing. https://github.com/wentaozhu/AnatomyNet-for-anatomical-segmentation.git. |
Tasks | |
Published | 2018-08-15 |
URL | http://arxiv.org/abs/1808.05238v2 |
http://arxiv.org/pdf/1808.05238v2.pdf | |
PWC | https://paperswithcode.com/paper/anatomynet-deep-learning-for-fast-and-fully |
Repo | https://github.com/ginn24/ICE3050-41 |
Framework | tf |
Regularized Wasserstein Means for Aligning Distributional Data
Title | Regularized Wasserstein Means for Aligning Distributional Data |
Authors | Liang Mi, Wen Zhang, Yalin Wang |
Abstract | We propose to align distributional data from the perspective of Wasserstein means. We raise the problem of regularizing Wasserstein means and propose several terms tailored to tackle different problems. Our formulation is based on the variational transportation to distribute a sparse discrete measure into the target domain. The resulting sparse representation well captures the desired property of the domain while reducing the mapping cost. We demonstrate the scalability and robustness of our method with examples in domain adaptation, point set registration, and skeleton layout. |
Tasks | Domain Adaptation |
Published | 2018-12-02 |
URL | https://arxiv.org/abs/1812.00338v2 |
https://arxiv.org/pdf/1812.00338v2.pdf | |
PWC | https://paperswithcode.com/paper/regularized-wasserstein-means-based-on |
Repo | https://github.com/icemiliang/pyvot |
Framework | pytorch |
Metric on Nonlinear Dynamical Systems with Perron-Frobenius Operators
Title | Metric on Nonlinear Dynamical Systems with Perron-Frobenius Operators |
Authors | Isao Ishikawa, Keisuke Fujii, Masahiro Ikeda, Yuka Hashimoto, Yoshinobu Kawahara |
Abstract | The development of a metric for structural data is a long-term problem in pattern recognition and machine learning. In this paper, we develop a general metric for comparing nonlinear dynamical systems that is defined with Perron-Frobenius operators in reproducing kernel Hilbert spaces. Our metric includes the existing fundamental metrics for dynamical systems, which are basically defined with principal angles between some appropriately-chosen subspaces, as its special cases. We also describe the estimation of our metric from finite data. We empirically illustrate our metric with an example of rotation dynamics in a unit disk in a complex plane, and evaluate the performance with real-world time-series data. |
Tasks | Time Series |
Published | 2018-05-31 |
URL | http://arxiv.org/abs/1805.12324v2 |
http://arxiv.org/pdf/1805.12324v2.pdf | |
PWC | https://paperswithcode.com/paper/metric-on-nonlinear-dynamical-systems-with |
Repo | https://github.com/keisuke198619/metricNLDS |
Framework | none |
Recurrent Neural Networks for Fuzz Testing Web Browsers
Title | Recurrent Neural Networks for Fuzz Testing Web Browsers |
Authors | Martin Sablotny, Bjørn Sand Jensen, Chris W. Johnson |
Abstract | Generation-based fuzzing is a software testing approach which is able to discover different types of bugs and vulnerabilities in software. It is, however, known to be very time consuming to design and fine tune classical fuzzers to achieve acceptable coverage, even for small-scale software systems. To address this issue, we investigate a machine learning-based approach to fuzz testing in which we outline a family of test-case generators based on Recurrent Neural Networks (RNNs) and train those on readily available datasets with a minimum of human fine tuning. The proposed generators do, in contrast to previous work, not rely on heuristic sampling strategies but principled sampling from the predictive distributions. We provide a detailed analysis to demonstrate the characteristics and efficacy of the proposed generators in a challenging web browser testing scenario. The empirical results show that the RNN-based generators are able to provide better coverage than a mutation based method and are able to discover paths not discovered by a classical fuzzer. Our results supplement findings in other domains suggesting that generation based fuzzing with RNNs is a viable route to better software quality conditioned on the use of a suitable model selection/analysis procedure. |
Tasks | Model Selection |
Published | 2018-12-12 |
URL | http://arxiv.org/abs/1812.04852v1 |
http://arxiv.org/pdf/1812.04852v1.pdf | |
PWC | https://paperswithcode.com/paper/recurrent-neural-networks-for-fuzz-testing |
Repo | https://github.com/susperius/icisc_rnnfuzz |
Framework | tf |
Parsimonious Bayesian deep networks
Title | Parsimonious Bayesian deep networks |
Authors | Mingyuan Zhou |
Abstract | Combining Bayesian nonparametrics and a forward model selection strategy, we construct parsimonious Bayesian deep networks (PBDNs) that infer capacity-regularized network architectures from the data and require neither cross-validation nor fine-tuning when training the model. One of the two essential components of a PBDN is the development of a special infinite-wide single-hidden-layer neural network, whose number of active hidden units can be inferred from the data. The other one is the construction of a greedy layer-wise learning algorithm that uses a forward model selection criterion to determine when to stop adding another hidden layer. We develop both Gibbs sampling and stochastic gradient descent based maximum a posteriori inference for PBDNs, providing state-of-the-art classification accuracy and interpretable data subtypes near the decision boundaries, while maintaining low computational complexity for out-of-sample prediction. |
Tasks | Model Selection |
Published | 2018-05-22 |
URL | http://arxiv.org/abs/1805.08719v3 |
http://arxiv.org/pdf/1805.08719v3.pdf | |
PWC | https://paperswithcode.com/paper/parsimonious-bayesian-deep-networks |
Repo | https://github.com/ethanhezhao/WEDTM |
Framework | none |
Latent Fingerprint Recognition: Role of Texture Template
Title | Latent Fingerprint Recognition: Role of Texture Template |
Authors | Kai Cao, Anil K. Jain |
Abstract | We propose a texture template approach, consisting of a set of virtual minutiae, to improve the overall latent fingerprint recognition accuracy. To compensate for the lack of sufficient number of minutiae in poor quality latent prints, we generate a set of virtual minutiae. However, due to a large number of these regularly placed virtual minutiae, texture based template matching has a large computational requirement compared to matching true minutiae templates. To improve both the accuracy and efficiency of the texture template matching, we investigate: i) both original and enhanced fingerprint patches for training convolutional neural networks (ConvNets) to improve the distinctiveness of descriptors associated with each virtual minutiae, ii) smaller patches around virtual minutiae and a fast ConvNet architecture to speed up descriptor extraction, iii) reduce the descriptor length, iv) a modified hierarchical graph matching strategy to improve the matching speed, and v) extraction of multiple texture templates to boost the performance. Experiments on NIST SD27 latent database show that the above strategies can improve the matching speed from 11 ms (24 threads) per comparison (between a latent and a reference print) to only 7.7 ms (single thread) per comparison while improving the rank-1 accuracy by 8.9% against 10K gallery. |
Tasks | Graph Matching |
Published | 2018-04-27 |
URL | http://arxiv.org/abs/1804.10337v1 |
http://arxiv.org/pdf/1804.10337v1.pdf | |
PWC | https://paperswithcode.com/paper/latent-fingerprint-recognition-role-of |
Repo | https://github.com/prip-lab/MSU-LatentAFIS |
Framework | pytorch |
Pseudo-Random Number Generation using Generative Adversarial Networks
Title | Pseudo-Random Number Generation using Generative Adversarial Networks |
Authors | Marcello De Bernardi, MHR Khouzani, Pasquale Malacaria |
Abstract | Pseudo-random number generators (PRNG) are a fundamental element of many security algorithms. We introduce a novel approach to their implementation, by proposing the use of generative adversarial networks (GAN) to train a neural network to behave as a PRNG. Furthermore, we showcase a number of interesting modifications to the standard GAN architecture. The most significant is partially concealing the output of the GAN’s generator, and training the adversary to discover a mapping from the overt part to the concealed part. The generator therefore learns to produce values the adversary cannot predict, rather than to approximate an explicit reference distribution. We demonstrate that a GAN can effectively train even a small feed-forward fully connected neural network to produce pseudo-random number sequences with good statistical properties. At best, subjected to the NIST test suite, the trained generator passed around 99% of test instances and 98% of overall tests, outperforming a number of standard non-cryptographic PRNGs. |
Tasks | |
Published | 2018-09-30 |
URL | http://arxiv.org/abs/1810.00378v1 |
http://arxiv.org/pdf/1810.00378v1.pdf | |
PWC | https://paperswithcode.com/paper/pseudo-random-number-generation-using |
Repo | https://github.com/marcellodebernardi/adversarial-csprng |
Framework | tf |
Provably Efficient Maximum Entropy Exploration
Title | Provably Efficient Maximum Entropy Exploration |
Authors | Elad Hazan, Sham M. Kakade, Karan Singh, Abby Van Soest |
Abstract | Suppose an agent is in a (possibly unknown) Markov Decision Process in the absence of a reward signal, what might we hope that an agent can efficiently learn to do? This work studies a broad class of objectives that are defined solely as functions of the state-visitation frequencies that are induced by how the agent behaves. For example, one natural, intrinsically defined, objective problem is for the agent to learn a policy which induces a distribution over state space that is as uniform as possible, which can be measured in an entropic sense. We provide an efficient algorithm to optimize such such intrinsically defined objectives, when given access to a black box planning oracle (which is robust to function approximation). Furthermore, when restricted to the tabular setting where we have sample based access to the MDP, our proposed algorithm is provably efficient, both in terms of its sample and computational complexities. Key to our algorithmic methodology is utilizing the conditional gradient method (a.k.a. the Frank-Wolfe algorithm) which utilizes an approximate MDP solver. |
Tasks | |
Published | 2018-12-06 |
URL | http://arxiv.org/abs/1812.02690v2 |
http://arxiv.org/pdf/1812.02690v2.pdf | |
PWC | https://paperswithcode.com/paper/provably-efficient-maximum-entropy |
Repo | https://github.com/abbyvansoest/maxent |
Framework | tf |
Isolating Sources of Disentanglement in Variational Autoencoders
Title | Isolating Sources of Disentanglement in Variational Autoencoders |
Authors | Ricky T. Q. Chen, Xuechen Li, Roger Grosse, David Duvenaud |
Abstract | We decompose the evidence lower bound to show the existence of a term measuring the total correlation between latent variables. We use this to motivate our $\beta$-TCVAE (Total Correlation Variational Autoencoder), a refinement of the state-of-the-art $\beta$-VAE objective for learning disentangled representations, requiring no additional hyperparameters during training. We further propose a principled classifier-free measure of disentanglement called the mutual information gap (MIG). We perform extensive quantitative and qualitative experiments, in both restricted and non-restricted settings, and show a strong relation between total correlation and disentanglement, when the latent variables model is trained using our framework. |
Tasks | |
Published | 2018-02-14 |
URL | http://arxiv.org/abs/1802.04942v5 |
http://arxiv.org/pdf/1802.04942v5.pdf | |
PWC | https://paperswithcode.com/paper/isolating-sources-of-disentanglement-in |
Repo | https://github.com/rtqichen/beta-tcvae |
Framework | pytorch |
VoiceFilter: Targeted Voice Separation by Speaker-Conditioned Spectrogram Masking
Title | VoiceFilter: Targeted Voice Separation by Speaker-Conditioned Spectrogram Masking |
Authors | Quan Wang, Hannah Muckenhirn, Kevin Wilson, Prashant Sridhar, Zelin Wu, John Hershey, Rif A. Saurous, Ron J. Weiss, Ye Jia, Ignacio Lopez Moreno |
Abstract | In this paper, we present a novel system that separates the voice of a target speaker from multi-speaker signals, by making use of a reference signal from the target speaker. We achieve this by training two separate neural networks: (1) A speaker recognition network that produces speaker-discriminative embeddings; (2) A spectrogram masking network that takes both noisy spectrogram and speaker embedding as input, and produces a mask. Our system significantly reduces the speech recognition WER on multi-speaker signals, with minimal WER degradation on single-speaker signals. |
Tasks | Speaker Recognition, Speaker Separation, Speech Enhancement, Speech Recognition |
Published | 2018-10-11 |
URL | https://arxiv.org/abs/1810.04826v6 |
https://arxiv.org/pdf/1810.04826v6.pdf | |
PWC | https://paperswithcode.com/paper/voicefilter-targeted-voice-separation-by |
Repo | https://github.com/mindslab-ai/voicefilter |
Framework | pytorch |
Generative Adversarial Networks for Image-to-Image Translation on Multi-Contrast MR Images - A Comparison of CycleGAN and UNIT
Title | Generative Adversarial Networks for Image-to-Image Translation on Multi-Contrast MR Images - A Comparison of CycleGAN and UNIT |
Authors | Per Welander, Simon Karlsson, Anders Eklund |
Abstract | In medical imaging, a general problem is that it is costly and time consuming to collect high quality data from healthy and diseased subjects. Generative adversarial networks (GANs) is a deep learning method that has been developed for synthesizing data. GANs can thereby be used to generate more realistic training data, to improve classification performance of machine learning algorithms. Another application of GANs is image-to-image translations, e.g. generating magnetic resonance (MR) images from computed tomography (CT) images, which can be used to obtain multimodal datasets from a single modality. Here, we evaluate two unsupervised GAN models (CycleGAN and UNIT) for image-to-image translation of T1- and T2-weighted MR images, by comparing generated synthetic MR images to ground truth images. We also evaluate two supervised models; a modification of CycleGAN and a pure generator model. A small perceptual study was also performed to evaluate how visually realistic the synthesized images are. It is shown that the implemented GAN models can synthesize visually realistic MR images (incorrectly labeled as real by a human). It is also shown that models producing more visually realistic synthetic images not necessarily have better quantitative error measurements, when compared to ground truth data. Code is available at https://github.com/simontomaskarlsson/GAN-MRI |
Tasks | Computed Tomography (CT), Image-to-Image Translation, Medical Image Generation |
Published | 2018-06-20 |
URL | http://arxiv.org/abs/1806.07777v1 |
http://arxiv.org/pdf/1806.07777v1.pdf | |
PWC | https://paperswithcode.com/paper/generative-adversarial-networks-for-image-to |
Repo | https://github.com/Siddhartha24795/Medical-Image-Synthesis |
Framework | tf |
Image-to-Image Translation via Group-wise Deep Whitening-and-Coloring Transformation
Title | Image-to-Image Translation via Group-wise Deep Whitening-and-Coloring Transformation |
Authors | Wonwoong Cho, Sungha Choi, David Keetae Park, Inkyu Shin, Jaegul Choo |
Abstract | Recently, unsupervised exemplar-based image-to-image translation, conditioned on a given exemplar without the paired data, has accomplished substantial advancements. In order to transfer the information from an exemplar to an input image, existing methods often use a normalization technique, e.g., adaptive instance normalization, that controls the channel-wise statistics of an input activation map at a particular layer, such as the mean and the variance. Meanwhile, style transfer approaches similar task to image translation by nature, demonstrated superior performance by using the higher-order statistics such as covariance among channels in representing a style. In detail, it works via whitening (given a zero-mean input feature, transforming its covariance matrix into the identity). followed by coloring (changing the covariance matrix of the whitened feature to those of the style feature). However, applying this approach in image translation is computationally intensive and error-prone due to the expensive time complexity and its non-trivial backpropagation. In response, this paper proposes an end-to-end approach tailored for image translation that efficiently approximates this transformation with our novel regularization methods. We further extend our approach to a group-wise form for memory and time efficiency as well as image quality. Extensive qualitative and quantitative experiments demonstrate that our proposed method is fast, both in training and inference, and highly effective in reflecting the style of an exemplar. Finally, our code is available at https://github.com/WonwoongCho/GDWCT. |
Tasks | Image-to-Image Translation, Style Transfer |
Published | 2018-12-24 |
URL | https://arxiv.org/abs/1812.09912v2 |
https://arxiv.org/pdf/1812.09912v2.pdf | |
PWC | https://paperswithcode.com/paper/image-to-image-translation-via-group-wise |
Repo | https://github.com/WonwoongCho/GDWCT |
Framework | pytorch |
Deep Optimisation: Solving Combinatorial Optimisation Problems using Deep Neural Networks
Title | Deep Optimisation: Solving Combinatorial Optimisation Problems using Deep Neural Networks |
Authors | J. R. Caldwell, R. A. Watson, C. Thies, J. D. Knowles |
Abstract | Deep Optimisation (DO) combines evolutionary search with Deep Neural Networks (DNNs) in a novel way - not for optimising a learning algorithm, but for finding a solution to an optimisation problem. Deep learning has been successfully applied to classification, regression, decision and generative tasks and in this paper we extend its application to solving optimisation problems. Model Building Optimisation Algorithms (MBOAs), a branch of evolutionary algorithms, have been successful in combining machine learning methods and evolutionary search but, until now, they have not utilised DNNs. DO is the first algorithm to use a DNN to learn and exploit the problem structure to adapt the variation operator (changing the neighbourhood structure of the search process). We demonstrate the performance of DO using two theoretical optimisation problems within the MAXSAT class. The Hierarchical Transformation Optimisation Problem (HTOP) has controllable deep structure that provides a clear evaluation of how DO works and why using a layerwise technique is essential for learning and exploiting problem structure. The Parity Modular Constraint Problem (MCparity) is a simplistic example of a problem containing higher-order dependencies (greater than pairwise) which DO can solve and state of the art MBOAs cannot. Further, we show that DO can exploit deep structure in TSP instances. Together these results show that there exists problems that DO can find and exploit deep problem structure that other algorithms cannot. Making this connection between DNNs and optimisation allows for the utilisation of advanced tools applicable to DNNs that current MBOAs are unable to use. |
Tasks | |
Published | 2018-11-02 |
URL | http://arxiv.org/abs/1811.00784v1 |
http://arxiv.org/pdf/1811.00784v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-optimisation-solving-combinatorial |
Repo | https://github.com/JimC12/Deep-Optimisation |
Framework | tf |
Efficient Contextualized Representation: Language Model Pruning for Sequence Labeling
Title | Efficient Contextualized Representation: Language Model Pruning for Sequence Labeling |
Authors | Liyuan Liu, Xiang Ren, Jingbo Shang, Jian Peng, Jiawei Han |
Abstract | Many efforts have been made to facilitate natural language processing tasks with pre-trained language models (LMs), and brought significant improvements to various applications. To fully leverage the nearly unlimited corpora and capture linguistic information of multifarious levels, large-size LMs are required; but for a specific task, only parts of these information are useful. Such large-sized LMs, even in the inference stage, may cause heavy computation workloads, making them too time-consuming for large-scale applications. Here we propose to compress bulky LMs while preserving useful information with regard to a specific task. As different layers of the model keep different information, we develop a layer selection method for model pruning using sparsity-inducing regularization. By introducing the dense connectivity, we can detach any layer without affecting others, and stretch shallow and wide LMs to be deep and narrow. In model training, LMs are learned with layer-wise dropouts for better robustness. Experiments on two benchmark datasets demonstrate the effectiveness of our method. |
Tasks | Language Modelling, Named Entity Recognition |
Published | 2018-04-20 |
URL | http://arxiv.org/abs/1804.07827v2 |
http://arxiv.org/pdf/1804.07827v2.pdf | |
PWC | https://paperswithcode.com/paper/efficient-contextualized-representation |
Repo | https://github.com/LiyuanLucasLiu/LD-Net |
Framework | pytorch |