Paper Group ANR 234
Blind Geometric Distortion Correction on Images Through Deep Learning. Improving One-shot NAS by Suppressing the Posterior Fading. Semi-Supervised Semantic Matching. Synthesis of Realistic ECG using Generative Adversarial Networks. Towards Content Transfer through Grounded Text Generation. An Update on Machine Learning in Neuro-oncology Diagnostics …
Blind Geometric Distortion Correction on Images Through Deep Learning
Title | Blind Geometric Distortion Correction on Images Through Deep Learning |
Authors | Xiaoyu Li, Bo Zhang, Pedro V. Sander, Jing Liao |
Abstract | We propose the first general framework to automatically correct different types of geometric distortion in a single input image. Our proposed method employs convolutional neural networks (CNNs) trained by using a large synthetic distortion dataset to predict the displacement field between distorted images and corrected images. A model fitting method uses the CNN output to estimate the distortion parameters, achieving a more accurate prediction. The final corrected image is generated based on the predicted flow using an efficient, high-quality resampling method. Experimental results demonstrate that our algorithm outperforms traditional correction methods, and allows for interesting applications such as distortion transfer, distortion exaggeration, and co-occurring distortion correction. |
Tasks | |
Published | 2019-09-08 |
URL | https://arxiv.org/abs/1909.03459v1 |
https://arxiv.org/pdf/1909.03459v1.pdf | |
PWC | https://paperswithcode.com/paper/blind-geometric-distortion-correction-on-1 |
Repo | |
Framework | |
Improving One-shot NAS by Suppressing the Posterior Fading
Title | Improving One-shot NAS by Suppressing the Posterior Fading |
Authors | Xiang Li, Chen Lin, Chuming Li, Ming Sun, Wei Wu, Junjie Yan, Wanli Ouyang |
Abstract | There is a growing interest in automated neural architecture search (NAS). To improve the efficiency of NAS, previous approaches adopt weight sharing method to force all models share the same set of weights. However, it has been observed that a model performing better with shared weights does not necessarily perform better when trained alone. In this paper, we analyse existing weight sharing one-shot NAS approaches from a Bayesian point of view and identify the posterior fading problem, which compromises the effectiveness of shared weights. To alleviate this problem, we present a practical approach to guide the parameter posterior towards its true distribution. Moreover, a hard latency constraint is introduced during the search so that the desired latency can be achieved. The resulted method, namely Posterior Convergent NAS (PC-NAS), achieves state-of-the-art performance under standard GPU latency constraint on ImageNet. In our small search space, our model PC-NAS-S attains 76.8 % top-1 accuracy, 2.1% higher than MobileNetV2 (1.4x) with the same latency. When adopted to the large search space, PC-NAS-L achieves 78.1 % top-1 accuracy within 11ms. The discovered architecture also transfers well to other computer vision applications such as object detection and person re-identification. |
Tasks | Neural Architecture Search, Object Detection, Person Re-Identification |
Published | 2019-10-06 |
URL | https://arxiv.org/abs/1910.02543v1 |
https://arxiv.org/pdf/1910.02543v1.pdf | |
PWC | https://paperswithcode.com/paper/improving-one-shot-nas-by-suppressing-the |
Repo | |
Framework | |
Semi-Supervised Semantic Matching
Title | Semi-Supervised Semantic Matching |
Authors | Zakaria Laskar, Juho Kannala |
Abstract | Convolutional neural networks (CNNs) have been successfully applied to solve the problem of correspondence estimation between semantically related images. Due to non-availability of large training datasets, existing methods resort to self-supervised or unsupervised training paradigm. In this paper we propose a semi-supervised learning framework that imposes cyclic consistency constraint on unlabeled image pairs. Together with the supervised loss the proposed model achieves state-of-the-art on a benchmark semantic matching dataset. |
Tasks | |
Published | 2019-01-24 |
URL | http://arxiv.org/abs/1901.08339v1 |
http://arxiv.org/pdf/1901.08339v1.pdf | |
PWC | https://paperswithcode.com/paper/semi-supervised-semantic-matching |
Repo | |
Framework | |
Synthesis of Realistic ECG using Generative Adversarial Networks
Title | Synthesis of Realistic ECG using Generative Adversarial Networks |
Authors | Anne Marie Delaney, Eoin Brophy, Tomas E. Ward |
Abstract | Access to medical data is highly restricted due to its sensitive nature, preventing communities from using this data for research or clinical training. Common methods of de-identification implemented to enable the sharing of data are sometimes inadequate to protect the individuals contained in the data. For our research, we investigate the ability of generative adversarial networks (GANs) to produce realistic medical time series data which can be used without concerns over privacy. The aim is to generate synthetic ECG signals representative of normal ECG waveforms. GANs have been used successfully to generate good quality synthetic time series and have been shown to prevent re-identification of individual records. In this work, a range of GAN architectures are developed to generate synthetic sine waves and synthetic ECG. Two evaluation metrics are then used to quantitatively assess how suitable the synthetic data is for real world applications such as clinical training and data analysis. Finally, we discuss the privacy concerns associated with sharing synthetic data produced by GANs and test their ability to withstand a simple membership inference attack. For the first time we both quantitatively and qualitatively demonstrate that GAN architecture can successfully generate time series signals that are not only structurally similar to the training sets but also diverse in nature across generated samples. We also report on their ability to withstand a simple membership inference attack, protecting the privacy of the training set. |
Tasks | Inference Attack, Time Series |
Published | 2019-09-19 |
URL | https://arxiv.org/abs/1909.09150v1 |
https://arxiv.org/pdf/1909.09150v1.pdf | |
PWC | https://paperswithcode.com/paper/synthesis-of-realistic-ecg-using-generative |
Repo | |
Framework | |
Towards Content Transfer through Grounded Text Generation
Title | Towards Content Transfer through Grounded Text Generation |
Authors | Shrimai Prabhumoye, Chris Quirk, Michel Galley |
Abstract | Recent work in neural generation has attracted significant interest in controlling the form of text, such as style, persona, and politeness. However, there has been less work on controlling neural text generation for content. This paper introduces the notion of Content Transfer for long-form text generation, where the task is to generate a next sentence in a document that both fits its context and is grounded in a content-rich external textual source such as a news story. Our experiments on Wikipedia data show significant improvements against competitive baselines. As another contribution of this paper, we release a benchmark dataset of 640k Wikipedia referenced sentences paired with the source articles to encourage exploration of this new task. |
Tasks | Text Generation |
Published | 2019-05-13 |
URL | https://arxiv.org/abs/1905.05293v1 |
https://arxiv.org/pdf/1905.05293v1.pdf | |
PWC | https://paperswithcode.com/paper/towards-content-transfer-through-grounded |
Repo | |
Framework | |
An Update on Machine Learning in Neuro-oncology Diagnostics
Title | An Update on Machine Learning in Neuro-oncology Diagnostics |
Authors | Thomas Booth |
Abstract | Imaging biomarkers in neuro-oncology are used for diagnosis, prognosis and treatment response monitoring. Magnetic resonance imaging is typically used throughout the patient pathway because routine structural imaging provides detailed anatomical and pathological information and advanced techniques provide additional physiological detail. Following image feature extraction, machine learning allows accurate classification in a variety of scenarios. Machine learning also enables image feature extraction de novo although the low prevalence of brain tumours makes such approaches challenging. Much research is applied to determining molecular profiles, histological tumour grade and prognosis at the time that patients first present with a brain tumour. Following treatment, differentiating a treatment response from a post-treatment related effect is clinically important and also an area of study. Most of the evidence is low level having been obtained retrospectively and in single centres. |
Tasks | |
Published | 2019-08-09 |
URL | https://arxiv.org/abs/1910.08157v1 |
https://arxiv.org/pdf/1910.08157v1.pdf | |
PWC | https://paperswithcode.com/paper/an-update-on-machine-learning-in-neuro |
Repo | |
Framework | |
Fully Automated Image De-fencing using Conditional Generative Adversarial Networks
Title | Fully Automated Image De-fencing using Conditional Generative Adversarial Networks |
Authors | Divyanshu Gupta, Shorya Jain, Utkarsh Tripathi, Pratik Chattopadhyay, Lipo Wang |
Abstract | Image de-fencing is one of the important aspects of recreational photography in which the objective is to remove the fence texture present in an image and generate an aesthetically pleasing version of the same image without the fence texture. In this paper, we aim to develop an automated and effective technique for fence removal and image reconstruction using conditional Generative Adversarial Networks (cGANs). These networks have been successfully applied in several domains of Computer Vision focusing on image generation and rendering. Our initial approach is based on a two-stage architecture involving two cGANs that generate the fence mask and the inpainted image, respectively. Training of these networks is carried out independently and, during evaluation, the input image is passed through the two generators in succession to obtain the de-fenced image. The results obtained from this approach are satisfactory, but the response time is long since the image has to pass through two sets of convolution layers. To reduce the response time, we propose a second approach involving only a single cGAN architecture that is trained using the ground-truth of fenced de-fenced image pairs along with the edge map of the fenced image produced by the Canny Filter. Incorporation of the edge map helps the network to precisely detect the edges present in the input image, and also imparts it an ability to carry out high quality de-fencing in an efficient manner, even in the presence of a fewer number of layers as compared to the two-stage network. Qualitative and quantitative experimental results reported in the manuscript reveal that the de-fenced images generated by the single-stage de-fencing network have similar visual quality to those produced by the two-stage network. Comparative performance analysis also emphasizes the effectiveness of our approach over state-of-the-art image de-fencing techniques. |
Tasks | Image Generation, Image Reconstruction |
Published | 2019-08-19 |
URL | https://arxiv.org/abs/1908.06837v1 |
https://arxiv.org/pdf/1908.06837v1.pdf | |
PWC | https://paperswithcode.com/paper/fully-automated-image-de-fencing-using |
Repo | |
Framework | |
Right-wing German Hate Speech on Twitter: Analysis and Automatic Detection
Title | Right-wing German Hate Speech on Twitter: Analysis and Automatic Detection |
Authors | Sylvia Jaki, Tom De Smedt |
Abstract | Discussion about the social network Twitter often concerns its role in political discourse, involving the question of when an expression of opinion becomes offensive, immoral, and/or illegal, and how to deal with it. Given the growing amount of offensive communication on the internet, there is a demand for new technology that can automatically detect hate speech, to assist content moderation by humans. This comes with new challenges, such as defining exactly what is free speech and what is illegal in a specific country, and knowing exactly what the linguistic characteristics of hate speech are. To shed light on the German situation, we analyzed over 50,000 right-wing German hate tweets posted between August 2017 and April 2018, at the time of the 2017 German federal elections, using both quantitative and qualitative methods. In this paper, we discuss the results of the analysis and demonstrate how the insights can be employed for the development of automatic detection systems. |
Tasks | |
Published | 2019-10-16 |
URL | https://arxiv.org/abs/1910.07518v1 |
https://arxiv.org/pdf/1910.07518v1.pdf | |
PWC | https://paperswithcode.com/paper/right-wing-german-hate-speech-on-twitter |
Repo | |
Framework | |
Identifying the Most Explainable Classifier
Title | Identifying the Most Explainable Classifier |
Authors | Brett Mullins |
Abstract | We introduce the notion of pointwise coverage to measure the explainability properties of machine learning classifiers. An explanation for a prediction is a definably simple region of the feature space sharing the same label as the prediction, and the coverage of an explanation measures its size or generalizability. With this notion of explanation, we investigate whether or not there is a natural characterization of the most explainable classifier. According with our intuitions, we prove that the binary linear classifier is uniquely the most explainable classifier up to negligible sets. |
Tasks | |
Published | 2019-10-18 |
URL | https://arxiv.org/abs/1910.08595v2 |
https://arxiv.org/pdf/1910.08595v2.pdf | |
PWC | https://paperswithcode.com/paper/identifying-the-most-explainable-classifier |
Repo | |
Framework | |
Four Things Everyone Should Know to Improve Batch Normalization
Title | Four Things Everyone Should Know to Improve Batch Normalization |
Authors | Cecilia Summers, Michael J. Dinneen |
Abstract | A key component of most neural network architectures is the use of normalization layers, such as Batch Normalization. Despite its common use and large utility in optimizing deep architectures, it has been challenging both to generically improve upon Batch Normalization and to understand the circumstances that lend themselves to other enhancements. In this paper, we identify four improvements to the generic form of Batch Normalization and the circumstances under which they work, yielding performance gains across all batch sizes while requiring no additional computation during training. These contributions include proposing a method for reasoning about the current example in inference normalization statistics, fixing a training vs. inference discrepancy; recognizing and validating the powerful regularization effect of Ghost Batch Normalization for small and medium batch sizes; examining the effect of weight decay regularization on the scaling and shifting parameters gamma and beta; and identifying a new normalization algorithm for very small batch sizes by combining the strengths of Batch and Group Normalization. We validate our results empirically on six datasets: CIFAR-100, SVHN, Caltech-256, Oxford Flowers-102, CUB-2011, and ImageNet. |
Tasks | |
Published | 2019-06-09 |
URL | https://arxiv.org/abs/1906.03548v2 |
https://arxiv.org/pdf/1906.03548v2.pdf | |
PWC | https://paperswithcode.com/paper/four-things-everyone-should-know-to-improve |
Repo | |
Framework | |
Context-Aware Embeddings for Automatic Art Analysis
Title | Context-Aware Embeddings for Automatic Art Analysis |
Authors | Noa Garcia, Benjamin Renoust, Yuta Nakashima |
Abstract | Automatic art analysis aims to classify and retrieve artistic representations from a collection of images by using computer vision and machine learning techniques. In this work, we propose to enhance visual representations from neural networks with contextual artistic information. Whereas visual representations are able to capture information about the content and the style of an artwork, our proposed context-aware embeddings additionally encode relationships between different artistic attributes, such as author, school, or historical period. We design two different approaches for using context in automatic art analysis. In the first one, contextual data is obtained through a multi-task learning model, in which several attributes are trained together to find visual relationships between elements. In the second approach, context is obtained through an art-specific knowledge graph, which encodes relationships between artistic attributes. An exhaustive evaluation of both of our models in several art analysis problems, such as author identification, type classification, or cross-modal retrieval, show that performance is improved by up to 7.3% in art classification and 37.24% in retrieval when context-aware embeddings are used. |
Tasks | Art Analysis, Cross-Modal Retrieval, Multi-Task Learning |
Published | 2019-04-10 |
URL | http://arxiv.org/abs/1904.04985v1 |
http://arxiv.org/pdf/1904.04985v1.pdf | |
PWC | https://paperswithcode.com/paper/context-aware-embeddings-for-automatic-art |
Repo | |
Framework | |
Some Limit Properties of Markov Chains Induced by Stochastic Recursive Algorithms
Title | Some Limit Properties of Markov Chains Induced by Stochastic Recursive Algorithms |
Authors | Abhishek Gupta, Gaurav Tendolkar, Hao Chen, Jianzong Pi |
Abstract | Recursive stochastic algorithms have gained significant attention in the recent past due to data driven applications. Examples include stochastic gradient descent for solving large-scale optimization problems and empirical dynamic programming algorithms for solving Markov decision problems. These recursive stochastic algorithms approximates certain contraction operators and can be viewed within the framework of iterated random maps. Accordingly, we consider iterated random maps over a Polish space that simulates a contraction operator over that Polish space. Assume that the iterated maps are indexed by $n$ such that as $n\rightarrow\infty$, each realization of the random map converges (in some sense) to the contraction map it is simulating. We show that starting from the same initial condition, the distribution of the random sequence generated by the iterated random maps converge weakly to the trajectory generated by the contraction operator. We further show that under certain conditions, the time average of the random sequence converge to the spatial mean of the invariant distribution. We then apply these results to logistic regression, empirical value iteration, empirical Q value iteration, and empirical relative value iteration for finite state finite action MDPs. |
Tasks | |
Published | 2019-04-24 |
URL | http://arxiv.org/abs/1904.10778v1 |
http://arxiv.org/pdf/1904.10778v1.pdf | |
PWC | https://paperswithcode.com/paper/some-limit-properties-of-markov-chains |
Repo | |
Framework | |
Attention-Wrapped Hierarchical BLSTMs for DDI Extraction
Title | Attention-Wrapped Hierarchical BLSTMs for DDI Extraction |
Authors | Vahab Mostafapour, Oğuz Dikenelli |
Abstract | Drug-Drug Interactions (DDIs) Extraction refers to the efforts to generate hand-made or automatic tools to extract embedded information from text and literature in the biomedical domain. Because of restrictions in hand-made efforts and their lower speed, Machine-Learning, or Deep-Learning approaches have become more popular for extracting DDIs. In this study, we propose a novel and generic Deep-Learning model which wraps Hierarchical Bidirectional LSTMs with two Attention Mechanisms that outperforms state-of-the-art models for DDIs Extraction, based on the DDIExtraction-2013 corpora. This model has obtained the macro F1-score of 0.785, and the precision of 0.80. |
Tasks | |
Published | 2019-07-31 |
URL | https://arxiv.org/abs/1907.13561v1 |
https://arxiv.org/pdf/1907.13561v1.pdf | |
PWC | https://paperswithcode.com/paper/attention-wrapped-hierarchical-blstms-for-ddi |
Repo | |
Framework | |
Learning Deep Generative Models with Short Run Inference Dynamics
Title | Learning Deep Generative Models with Short Run Inference Dynamics |
Authors | Erik Nijkamp, Bo Pang, Tian Han, Linqi Zhou, Song-Chun Zhu, Ying Nian Wu |
Abstract | This paper studies the fundamental problem of learning deep generative models that consist of one or more layers of latent variables organized in top-down architectures. Learning such a generative model requires inferring the latent variables for each training example based on the posterior distribution of these latent variables. The inference typically requires Markov chain Monte Caro (MCMC) that can be time consuming. In this paper, we propose to use short run inference dynamics guided by the log-posterior, such as finite-step gradient descent algorithm initialized from the prior distribution of the latent variables, as an approximate sampler of the posterior distribution, where the step size of the gradient descent dynamics is optimized by minimizing the Kullback-Leibler divergence between the distribution produced by the short run inference dynamics and the posterior distribution. Our experiments show that the proposed method outperforms variational auto-encoder (VAE) in terms of reconstruction error and synthesis quality. The advantage of the proposed method is that it is natural and automatic, even for models with multiple layers of latent variables. |
Tasks | |
Published | 2019-12-04 |
URL | https://arxiv.org/abs/1912.01909v3 |
https://arxiv.org/pdf/1912.01909v3.pdf | |
PWC | https://paperswithcode.com/paper/learning-deep-generative-models-with-short |
Repo | |
Framework | |
Preservation of Anomalous Subgroups On Machine Learning Transformed Data
Title | Preservation of Anomalous Subgroups On Machine Learning Transformed Data |
Authors | Samuel C. Maina, Reginald E. Bryant, William O. Goal, Robert-Florian Samoilescu, Kush R. Varshney, Komminist Weldemariam |
Abstract | In this paper, we investigate the effect of machine learning based anonymization on anomalous subgroup preservation. In particular, we train a binary classifier to discover the most anomalous subgroup in a dataset by maximizing the bias between the group’s predicted odds ratio from the model and observed odds ratio from the data. We then perform anonymization using a variational autoencoder (VAE) to synthesize an entirely new dataset that would ideally be drawn from the distribution of the original data. We repeat the anomalous subgroup discovery task on the new data and compare it to what was identified pre-anonymization. We evaluated our approach using publicly available datasets from the financial industry. Our evaluation confirmed that the approach was able to produce synthetic datasets that preserved a high level of subgroup differentiation as identified initially in the original dataset. Such a distinction was maintained while having distinctly different records between the synthetic and original dataset. Finally, we packed the above end to end process into what we call Utility Guaranteed Deep Privacy (UGDP) system. UGDP can be easily extended to onboard alternative generative approaches such as GANs to synthesize tabular data. |
Tasks | |
Published | 2019-11-09 |
URL | https://arxiv.org/abs/1911.03674v1 |
https://arxiv.org/pdf/1911.03674v1.pdf | |
PWC | https://paperswithcode.com/paper/preservation-of-anomalous-subgroups-on |
Repo | |
Framework | |