Paper Group AWR 167
FeUdal Networks for Hierarchical Reinforcement Learning. Predicting SMT Solver Performance for Software Verification. Paying Attention to Multi-Word Expressions in Neural Machine Translation. Low Dose CT Image Denoising Using a Generative Adversarial Network with Wasserstein Distance and Perceptual Loss. Learning Dilation Factors for Semantic Segme …
FeUdal Networks for Hierarchical Reinforcement Learning
Title | FeUdal Networks for Hierarchical Reinforcement Learning |
Authors | Alexander Sasha Vezhnevets, Simon Osindero, Tom Schaul, Nicolas Heess, Max Jaderberg, David Silver, Koray Kavukcuoglu |
Abstract | We introduce FeUdal Networks (FuNs): a novel architecture for hierarchical reinforcement learning. Our approach is inspired by the feudal reinforcement learning proposal of Dayan and Hinton, and gains power and efficacy by decoupling end-to-end learning across multiple levels – allowing it to utilise different resolutions of time. Our framework employs a Manager module and a Worker module. The Manager operates at a lower temporal resolution and sets abstract goals which are conveyed to and enacted by the Worker. The Worker generates primitive actions at every tick of the environment. The decoupled structure of FuN conveys several benefits – in addition to facilitating very long timescale credit assignment it also encourages the emergence of sub-policies associated with different goals set by the Manager. These properties allow FuN to dramatically outperform a strong baseline agent on tasks that involve long-term credit assignment or memorisation. We demonstrate the performance of our proposed system on a range of tasks from the ATARI suite and also from a 3D DeepMind Lab environment. |
Tasks | Hierarchical Reinforcement Learning |
Published | 2017-03-03 |
URL | http://arxiv.org/abs/1703.01161v2 |
http://arxiv.org/pdf/1703.01161v2.pdf | |
PWC | https://paperswithcode.com/paper/feudal-networks-for-hierarchical |
Repo | https://github.com/danielpalen/pysc2-rl-agents |
Framework | tf |
Predicting SMT Solver Performance for Software Verification
Title | Predicting SMT Solver Performance for Software Verification |
Authors | Andrew Healy, Rosemary Monahan, James F. Power |
Abstract | The Why3 IDE and verification system facilitates the use of a wide range of Satisfiability Modulo Theories (SMT) solvers through a driver-based architecture. We present Where4: a portfolio-based approach to discharge Why3 proof obligations. We use data analysis and machine learning techniques on static metrics derived from program source code. Our approach benefits software engineers by providing a single utility to delegate proof obligations to the solvers most likely to return a useful result. It does this in a time-efficient way using existing Why3 and solver installations - without requiring low-level knowledge about SMT solver operation from the user. |
Tasks | |
Published | 2017-01-30 |
URL | http://arxiv.org/abs/1701.08466v1 |
http://arxiv.org/pdf/1701.08466v1.pdf | |
PWC | https://paperswithcode.com/paper/predicting-smt-solver-performance-for |
Repo | https://github.com/ahealy19/F-IDE-2016 |
Framework | none |
Paying Attention to Multi-Word Expressions in Neural Machine Translation
Title | Paying Attention to Multi-Word Expressions in Neural Machine Translation |
Authors | Matīss Rikters, Ondřej Bojar |
Abstract | Processing of multi-word expressions (MWEs) is a known problem for any natural language processing task. Even neural machine translation (NMT) struggles to overcome it. This paper presents results of experiments on investigating NMT attention allocation to the MWEs and improving automated translation of sentences that contain MWEs in English->Latvian and English->Czech NMT systems. Two improvement strategies were explored -(1) bilingual pairs of automatically extracted MWE candidates were added to the parallel corpus used to train the NMT system, and (2) full sentences containing the automatically extracted MWE candidates were added to the parallel corpus. Both approaches allowed to increase automated evaluation results. The best result - 0.99 BLEU point increase - has been reached with the first approach, while with the second approach minimal improvements achieved. We also provide open-source software and tools used for MWE extraction and alignment inspection. |
Tasks | Machine Translation |
Published | 2017-10-17 |
URL | https://arxiv.org/abs/1710.06313v2 |
https://arxiv.org/pdf/1710.06313v2.pdf | |
PWC | https://paperswithcode.com/paper/paying-attention-to-multi-word-expressions-in |
Repo | https://github.com/M4t1ss/MWE-Tools |
Framework | none |
Low Dose CT Image Denoising Using a Generative Adversarial Network with Wasserstein Distance and Perceptual Loss
Title | Low Dose CT Image Denoising Using a Generative Adversarial Network with Wasserstein Distance and Perceptual Loss |
Authors | Qingsong Yang, Pingkun Yan, Yanbo Zhang, Hengyong Yu, Yongyi Shi, Xuanqin Mou, Mannudeep K. Kalra, Ge Wang |
Abstract | In this paper, we introduce a new CT image denoising method based on the generative adversarial network (GAN) with Wasserstein distance and perceptual similarity. The Wasserstein distance is a key concept of the optimal transform theory, and promises to improve the performance of the GAN. The perceptual loss compares the perceptual features of a denoised output against those of the ground truth in an established feature space, while the GAN helps migrate the data noise distribution from strong to weak. Therefore, our proposed method transfers our knowledge of visual perception to the image denoising task, is capable of not only reducing the image noise level but also keeping the critical information at the same time. Promising results have been obtained in our experiments with clinical CT images. |
Tasks | Denoising, Image Denoising |
Published | 2017-08-03 |
URL | http://arxiv.org/abs/1708.00961v2 |
http://arxiv.org/pdf/1708.00961v2.pdf | |
PWC | https://paperswithcode.com/paper/low-dose-ct-image-denoising-using-a |
Repo | https://github.com/daintlab/ct-denoising |
Framework | tf |
Learning Dilation Factors for Semantic Segmentation of Street Scenes
Title | Learning Dilation Factors for Semantic Segmentation of Street Scenes |
Authors | Yang He, Margret Keuper, Bernt Schiele, Mario Fritz |
Abstract | Contextual information is crucial for semantic segmentation. However, finding the optimal trade-off between keeping desired fine details and at the same time providing sufficiently large receptive fields is non trivial. This is even more so, when objects or classes present in an image significantly vary in size. Dilated convolutions have proven valuable for semantic segmentation, because they allow to increase the size of the receptive field without sacrificing image resolution. However, in current state-of-the-art methods, dilation parameters are hand-tuned and fixed. In this paper, we present an approach for learning dilation parameters adaptively per channel, consistently improving semantic segmentation results on street-scene datasets like Cityscapes and Camvid. |
Tasks | Semantic Segmentation |
Published | 2017-09-06 |
URL | http://arxiv.org/abs/1709.01956v1 |
http://arxiv.org/pdf/1709.01956v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-dilation-factors-for-semantic |
Repo | https://github.com/SSAW14/LearnableDilationNetwork |
Framework | none |
Towards Building a Knowledge Base of Monetary Transactions from a News Collection
Title | Towards Building a Knowledge Base of Monetary Transactions from a News Collection |
Authors | Jan R. Benetka, Krisztian Balog, Kjetil Nørvåg |
Abstract | We address the problem of extracting structured representations of economic events from a large corpus of news articles, using a combination of natural language processing and machine learning techniques. The developed techniques allow for semi-automatic population of a financial knowledge base, which, in turn, may be used to support a range of data mining and exploration tasks. The key challenge we face in this domain is that the same event is often reported multiple times, with varying correctness of details. We address this challenge by first collecting all information pertinent to a given event from the entire corpus, then considering all possible representations of the event, and finally, using a supervised learning method, to rank these representations by the associated confidence scores. A main innovative element of our approach is that it jointly extracts and stores all attributes of the event as a single representation (quintuple). Using a purpose-built test set we demonstrate that our supervised learning approach can achieve 25% improvement in F1-score over baseline methods that consider the earliest, the latest or the most frequent reporting of the event. |
Tasks | |
Published | 2017-09-18 |
URL | http://arxiv.org/abs/1709.05743v1 |
http://arxiv.org/pdf/1709.05743v1.pdf | |
PWC | https://paperswithcode.com/paper/towards-building-a-knowledge-base-of-monetary |
Repo | https://github.com/benetka/kbmt |
Framework | none |
CLTune: A Generic Auto-Tuner for OpenCL Kernels
Title | CLTune: A Generic Auto-Tuner for OpenCL Kernels |
Authors | Cedric Nugteren, Valeriu Codreanu |
Abstract | This work presents CLTune, an auto-tuner for OpenCL kernels. It evaluates and tunes kernel performance of a generic, user-defined search space of possible parameter-value combinations. Example parameters include the OpenCL workgroup size, vector data-types, tile sizes, and loop unrolling factors. CLTune can be used in the following scenarios: 1) when there are too many tunable parameters to explore manually, 2) when performance portability across OpenCL devices is desired, or 3) when the optimal parameters change based on input argument values (e.g. matrix dimensions). The auto-tuner is generic, easy to use, open-source, and supports multiple search strategies including simulated annealing and particle swarm optimisation. CLTune is evaluated on two GPU case-studies inspired by the recent successes in deep learning: 2D convolution and matrix-multiplication (GEMM). For 2D convolution, we demonstrate the need for auto-tuning by optimizing for different filter sizes, achieving performance on-par or better than the state-of-the-art. For matrix-multiplication, we use CLTune to explore a parameter space of more than two-hundred thousand configurations, we show the need for device-specific tuning, and outperform the clBLAS library on NVIDIA, AMD and Intel GPUs. |
Tasks | |
Published | 2017-03-19 |
URL | http://arxiv.org/abs/1703.06503v1 |
http://arxiv.org/pdf/1703.06503v1.pdf | |
PWC | https://paperswithcode.com/paper/cltune-a-generic-auto-tuner-for-opencl |
Repo | https://github.com/CNugteren/CLBlast |
Framework | none |
Unrestricted Facial Geometry Reconstruction Using Image-to-Image Translation
Title | Unrestricted Facial Geometry Reconstruction Using Image-to-Image Translation |
Authors | Matan Sela, Elad Richardson, Ron Kimmel |
Abstract | It has been recently shown that neural networks can recover the geometric structure of a face from a single given image. A common denominator of most existing face geometry reconstruction methods is the restriction of the solution space to some low-dimensional subspace. While such a model significantly simplifies the reconstruction problem, it is inherently limited in its expressiveness. As an alternative, we propose an Image-to-Image translation network that jointly maps the input image to a depth image and a facial correspondence map. This explicit pixel-based mapping can then be utilized to provide high quality reconstructions of diverse faces under extreme expressions, using a purely geometric refinement process. In the spirit of recent approaches, the network is trained only with synthetic data, and is then evaluated on in-the-wild facial images. Both qualitative and quantitative analyses demonstrate the accuracy and the robustness of our approach. |
Tasks | 3D Face Reconstruction, Image-to-Image Translation |
Published | 2017-03-29 |
URL | http://arxiv.org/abs/1703.10131v2 |
http://arxiv.org/pdf/1703.10131v2.pdf | |
PWC | https://paperswithcode.com/paper/unrestricted-facial-geometry-reconstruction |
Repo | https://github.com/matansel/pix2vertex |
Framework | torch |
Don’t Decay the Learning Rate, Increase the Batch Size
Title | Don’t Decay the Learning Rate, Increase the Batch Size |
Authors | Samuel L. Smith, Pieter-Jan Kindermans, Chris Ying, Quoc V. Le |
Abstract | It is common practice to decay the learning rate. Here we show one can usually obtain the same learning curve on both training and test sets by instead increasing the batch size during training. This procedure is successful for stochastic gradient descent (SGD), SGD with momentum, Nesterov momentum, and Adam. It reaches equivalent test accuracies after the same number of training epochs, but with fewer parameter updates, leading to greater parallelism and shorter training times. We can further reduce the number of parameter updates by increasing the learning rate $\epsilon$ and scaling the batch size $B \propto \epsilon$. Finally, one can increase the momentum coefficient $m$ and scale $B \propto 1/(1-m)$, although this tends to slightly reduce the test accuracy. Crucially, our techniques allow us to repurpose existing training schedules for large batch training with no hyper-parameter tuning. We train ResNet-50 on ImageNet to $76.1%$ validation accuracy in under 30 minutes. |
Tasks | |
Published | 2017-11-01 |
URL | http://arxiv.org/abs/1711.00489v2 |
http://arxiv.org/pdf/1711.00489v2.pdf | |
PWC | https://paperswithcode.com/paper/dont-decay-the-learning-rate-increase-the |
Repo | https://github.com/rbkim1990/capstone-age-estimation |
Framework | none |
ChemGAN challenge for drug discovery: can AI reproduce natural chemical diversity?
Title | ChemGAN challenge for drug discovery: can AI reproduce natural chemical diversity? |
Authors | Mostapha Benhenda |
Abstract | Generating molecules with desired chemical properties is important for drug discovery. The use of generative neural networks is promising for this task. However, from visual inspection, it often appears that generated samples lack diversity. In this paper, we quantify this internal chemical diversity, and we raise the following challenge: can a nontrivial AI model reproduce natural chemical diversity for desired molecules? To illustrate this question, we consider two generative models: a Reinforcement Learning model and the recently introduced ORGAN. Both fail at this challenge. We hope this challenge will stimulate research in this direction. |
Tasks | Drug Discovery |
Published | 2017-08-28 |
URL | http://arxiv.org/abs/1708.08227v3 |
http://arxiv.org/pdf/1708.08227v3.pdf | |
PWC | https://paperswithcode.com/paper/chemgan-challenge-for-drug-discovery-can-ai |
Repo | https://github.com/cool21th/ai_drug_discovery |
Framework | none |
Constant Size Molecular Descriptors For Use With Machine Learning
Title | Constant Size Molecular Descriptors For Use With Machine Learning |
Authors | Christopher R. Collins, Geoffrey J. Gordon, O. Anatole von Lilienfeld, David J. Yaron |
Abstract | A set of molecular descriptors whose length is independent of molecular size is developed for machine learning models that target thermodynamic and electronic properties of molecules. These features are evaluated by monitoring performance of kernel ridge regression models on well-studied data sets of small organic molecules. The features include connectivity counts, which require only the bonding pattern of the molecule, and encoded distances, which summarize distances between both bonded and non-bonded atoms and so require the full molecular geometry. In addition to having constant size, these features summarize information regarding the local environment of atoms and bonds, such that models can take advantage of similarities resulting from the presence of similar chemical fragments across molecules. Combining these two types of features leads to models whose performance is comparable to or better than the current state of the art. The features introduced here have the advantage of leading to models that may be trained on smaller molecules and then used successfully on larger molecules. |
Tasks | |
Published | 2017-01-23 |
URL | http://arxiv.org/abs/1701.06649v1 |
http://arxiv.org/pdf/1701.06649v1.pdf | |
PWC | https://paperswithcode.com/paper/constant-size-molecular-descriptors-for-use |
Repo | https://github.com/jsheng7/ANI1-qm7 |
Framework | none |
Broadband DOA estimation using Convolutional neural networks trained with noise signals
Title | Broadband DOA estimation using Convolutional neural networks trained with noise signals |
Authors | Soumitro Chakrabarty, Emanuël. A. P. Habets |
Abstract | A convolution neural network (CNN) based classification method for broadband DOA estimation is proposed, where the phase component of the short-time Fourier transform coefficients of the received microphone signals are directly fed into the CNN and the features required for DOA estimation are learnt during training. Since only the phase component of the input is used, the CNN can be trained with synthesized noise signals, thereby making the preparation of the training data set easier compared to using speech signals. Through experimental evaluation, the ability of the proposed noise trained CNN framework to generalize to speech sources is demonstrated. In addition, the robustness of the system to noise, small perturbations in microphone positions, as well as its ability to adapt to different acoustic conditions is investigated using experiments with simulated and real data. |
Tasks | |
Published | 2017-05-02 |
URL | http://arxiv.org/abs/1705.00919v2 |
http://arxiv.org/pdf/1705.00919v2.pdf | |
PWC | https://paperswithcode.com/paper/broadband-doa-estimation-using-convolutional |
Repo | https://github.com/Soumitro-Chakrabarty/Single-speaker-localization |
Framework | none |
Stochastic Generative Hashing
Title | Stochastic Generative Hashing |
Authors | Bo Dai, Ruiqi Guo, Sanjiv Kumar, Niao He, Le Song |
Abstract | Learning-based binary hashing has become a powerful paradigm for fast search and retrieval in massive databases. However, due to the requirement of discrete outputs for the hash functions, learning such functions is known to be very challenging. In addition, the objective functions adopted by existing hashing techniques are mostly chosen heuristically. In this paper, we propose a novel generative approach to learn hash functions through Minimum Description Length principle such that the learned hash codes maximally compress the dataset and can also be used to regenerate the inputs. We also develop an efficient learning algorithm based on the stochastic distributional gradient, which avoids the notorious difficulty caused by binary output constraints, to jointly optimize the parameters of the hash function and the associated generative model. Extensive experiments on a variety of large-scale datasets show that the proposed method achieves better retrieval results than the existing state-of-the-art methods. |
Tasks | |
Published | 2017-01-11 |
URL | http://arxiv.org/abs/1701.02815v2 |
http://arxiv.org/pdf/1701.02815v2.pdf | |
PWC | https://paperswithcode.com/paper/stochastic-generative-hashing |
Repo | https://github.com/doubling/Stochastic_Generative_Hashing |
Framework | tf |
Coarse-to-Fine Lifted MAP Inference in Computer Vision
Title | Coarse-to-Fine Lifted MAP Inference in Computer Vision |
Authors | Haroun Habeeb, Ankit Anand, Mausam, Parag Singla |
Abstract | There is a vast body of theoretical research on lifted inference in probabilistic graphical models (PGMs). However, few demonstrations exist where lifting is applied in conjunction with top of the line applied algorithms. We pursue the applicability of lifted inference for computer vision (CV), with the insight that a globally optimal (MAP) labeling will likely have the same label for two symmetric pixels. The success of our approach lies in efficiently handling a distinct unary potential on every node (pixel), typical of CV applications. This allows us to lift the large class of algorithms that model a CV problem via PGM inference. We propose a generic template for coarse-to-fine (C2F) inference in CV, which progressively refines an initial coarsely lifted PGM for varying quality-time trade-offs. We demonstrate the performance of C2F inference by developing lifted versions of two near state-of-the-art CV algorithms for stereo vision and interactive image segmentation. We find that, against flat algorithms, the lifted versions have a much superior anytime performance, without any loss in final solution quality. |
Tasks | Semantic Segmentation |
Published | 2017-07-22 |
URL | http://arxiv.org/abs/1707.07165v1 |
http://arxiv.org/pdf/1707.07165v1.pdf | |
PWC | https://paperswithcode.com/paper/coarse-to-fine-lifted-map-inference-in |
Repo | https://github.com/dair-iitd/c2fi4cv |
Framework | none |
Photorealistic Style Transfer with Screened Poisson Equation
Title | Photorealistic Style Transfer with Screened Poisson Equation |
Authors | Roey Mechrez, Eli Shechtman, Lihi Zelnik-Manor |
Abstract | Recent work has shown impressive success in transferring painterly style to images. These approaches, however, fall short of photorealistic style transfer. Even when both the input and reference images are photographs, the output still exhibits distortions reminiscent of a painting. In this paper we propose an approach that takes as input a stylized image and makes it more photorealistic. It relies on the Screened Poisson Equation, maintaining the fidelity of the stylized image while constraining the gradients to those of the original input image. Our method is fast, simple, fully automatic and shows positive progress in making a stylized image photorealistic. Our results exhibit finer details and are less prone to artifacts than the state-of-the-art. |
Tasks | Style Transfer |
Published | 2017-09-28 |
URL | http://arxiv.org/abs/1709.09828v1 |
http://arxiv.org/pdf/1709.09828v1.pdf | |
PWC | https://paperswithcode.com/paper/photorealistic-style-transfer-with-screened |
Repo | https://github.com/roimehrez/photorealism |
Framework | none |