Paper Group ANR 1779
AIM 2019 Challenge on Real-World Image Super-Resolution: Methods and Results. Frequency Separation for Real-World Super-Resolution. Training Set Effect on Super Resolution for Automated Target Recognition. Text Summarization in the Biomedical Domain. Summary Refinement through Denoising. STRASS: A Light and Effective Method for Extractive Summariza …
AIM 2019 Challenge on Real-World Image Super-Resolution: Methods and Results
Title | AIM 2019 Challenge on Real-World Image Super-Resolution: Methods and Results |
Authors | Andreas Lugmayr, Martin Danelljan, Radu Timofte, Manuel Fritsche, Shuhang Gu, Kuldeep Purohit, Praveen Kandula, Maitreya Suin, A N Rajagopalan, Nam Hyung Joon, Yu Seung Won, Guisik Kim, Dokyeong Kwon, Chih-Chung Hsu, Chia-Hsiang Lin, Yuanfei Huang, Xiaopeng Sun, Wen Lu, Jie Li, Xinbo Gao, Sefi Bell-Kligler |
Abstract | This paper reviews the AIM 2019 challenge on real world super-resolution. It focuses on the participating methods and final results. The challenge addresses the real world setting, where paired true high and low-resolution images are unavailable. For training, only one set of source input images is therefore provided in the challenge. In Track 1: Source Domain the aim is to super-resolve such images while preserving the low level image characteristics of the source input domain. In Track 2: Target Domain a set of high-quality images is also provided for training, that defines the output domain and desired quality of the super-resolved images. To allow for quantitative evaluation, the source input images in both tracks are constructed using artificial, but realistic, image degradations. The challenge is the first of its kind, aiming to advance the state-of-the-art and provide a standard benchmark for this newly emerging task. In total 7 teams competed in the final testing phase, demonstrating new and innovative solutions to the problem. |
Tasks | Image Super-Resolution, Super-Resolution |
Published | 2019-11-18 |
URL | https://arxiv.org/abs/1911.07783v2 |
https://arxiv.org/pdf/1911.07783v2.pdf | |
PWC | https://paperswithcode.com/paper/aim-2019-challenge-on-real-world-image-super |
Repo | |
Framework | |
Frequency Separation for Real-World Super-Resolution
Title | Frequency Separation for Real-World Super-Resolution |
Authors | Manuel Fritsche, Shuhang Gu, Radu Timofte |
Abstract | Most of the recent literature on image super-resolution (SR) assumes the availability of training data in the form of paired low resolution (LR) and high resolution (HR) images or the knowledge of the downgrading operator (usually bicubic downscaling). While the proposed methods perform well on standard benchmarks, they often fail to produce convincing results in real-world settings. This is because real-world images can be subject to corruptions such as sensor noise, which are severely altered by bicubic downscaling. Therefore, the models never see a real-world image during training, which limits their generalization capabilities. Moreover, it is cumbersome to collect paired LR and HR images in the same source domain. To address this problem, we propose DSGAN to introduce natural image characteristics in bicubically downscaled images. It can be trained in an unsupervised fashion on HR images, thereby generating LR images with the same characteristics as the original images. We then use the generated data to train a SR model, which greatly improves its performance on real-world images. Furthermore, we propose to separate the low and high image frequencies and treat them differently during training. Since the low frequencies are preserved by downsampling operations, we only require adversarial training to modify the high frequencies. This idea is applied to our DSGAN model as well as the SR model. We demonstrate the effectiveness of our method in several experiments through quantitative and qualitative analysis. Our solution is the winner of the AIM Challenge on Real World SR at ICCV 2019. |
Tasks | Image Super-Resolution, Super-Resolution |
Published | 2019-11-18 |
URL | https://arxiv.org/abs/1911.07850v1 |
https://arxiv.org/pdf/1911.07850v1.pdf | |
PWC | https://paperswithcode.com/paper/frequency-separation-for-real-world-super |
Repo | |
Framework | |
Training Set Effect on Super Resolution for Automated Target Recognition
Title | Training Set Effect on Super Resolution for Automated Target Recognition |
Authors | Matthew Ciolino, David Noever, Josh Kalin |
Abstract | Single Image Super Resolution (SISR) is the process of mapping a low-resolution image to a high resolution image. This inherently has applications in remote sensing as a way to increase the spatial resolution in satellite imagery. This suggests a possible improvement to automated target recognition in image classification and object detection. We explore the effect that different training sets have on SISR with the network, Super Resolution Generative Adversarial Network (SRGAN). We train 5 SRGANs on different land-use classes (e.g. agriculture, cities, ports) and test them on the same unseen dataset. We attempt to find the qualitative and quantitative differences in SISR, binary classification, and object detection performance. We find that curated training sets that contain objects in the test ontology perform better on both computer vision tasks while having a complex distribution of images allows object detection models to perform better. However, Super Resolution (SR) might not be beneficial to certain problems and will see a diminishing amount of returns for datasets that are closer to being solved. |
Tasks | Image Classification, Image Super-Resolution, Object Detection, Super-Resolution |
Published | 2019-10-29 |
URL | https://arxiv.org/abs/1911.07934v3 |
https://arxiv.org/pdf/1911.07934v3.pdf | |
PWC | https://paperswithcode.com/paper/training-set-affect-on-super-resolution-for |
Repo | |
Framework | |
Text Summarization in the Biomedical Domain
Title | Text Summarization in the Biomedical Domain |
Authors | Milad Moradi, Nasser Ghadiri |
Abstract | This chapter gives an overview of recent advances in the field of biomedical text summarization. Different types of challenges are introduced, and methods are discussed concerning the type of challenge that they address. Biomedical literature summarization is explored as a leading trend in the field, and some future lines of work are pointed out. Underlying methods of recent summarization systems are briefly explained and the most significant evaluation results are mentioned. The primary purpose of this chapter is to review the most significant research efforts made in the current decade toward new methods of biomedical text summarization. As the main parts of this chapter, current trends are discussed and new challenges are introduced. |
Tasks | Text Summarization |
Published | 2019-08-06 |
URL | https://arxiv.org/abs/1908.02285v1 |
https://arxiv.org/pdf/1908.02285v1.pdf | |
PWC | https://paperswithcode.com/paper/text-summarization-in-the-biomedical-domain |
Repo | |
Framework | |
Summary Refinement through Denoising
Title | Summary Refinement through Denoising |
Authors | Nikola I. Nikolov, Alessandro Calmanovici, Richard H. R. Hahnloser |
Abstract | We propose a simple method for post-processing the outputs of a text summarization system in order to refine its overall quality. Our approach is to train text-to-text rewriting models to correct information redundancy errors that may arise during summarization. We train on synthetically generated noisy summaries, testing three different types of noise that introduce out-of-context information within each summary. When applied on top of extractive and abstractive summarization baselines, our summary denoising models yield metric improvements while reducing redundancy. |
Tasks | Abstractive Text Summarization, Denoising, Text Summarization |
Published | 2019-07-25 |
URL | https://arxiv.org/abs/1907.10873v1 |
https://arxiv.org/pdf/1907.10873v1.pdf | |
PWC | https://paperswithcode.com/paper/summary-refinement-through-denoising |
Repo | |
Framework | |
STRASS: A Light and Effective Method for Extractive Summarization Based on Sentence Embeddings
Title | STRASS: A Light and Effective Method for Extractive Summarization Based on Sentence Embeddings |
Authors | Léo Bouscarrat, Antoine Bonnefoy, Thomas Peel, Cécile Pereira |
Abstract | This paper introduces STRASS: Summarization by TRAnsformation Selection and Scoring. It is an extractive text summarization method which leverages the semantic information in existing sentence embedding spaces. Our method creates an extractive summary by selecting the sentences with the closest embeddings to the document embedding. The model learns a transformation of the document embedding to minimize the similarity between the extractive summary and the ground truth summary. As the transformation is only composed of a dense layer, the training can be done on CPU, therefore, inexpensive. Moreover, inference time is short and linear according to the number of sentences. As a second contribution, we introduce the French CASS dataset, composed of judgments from the French Court of cassation and their corresponding summaries. On this dataset, our results show that our method performs similarly to the state of the art extractive methods with effective training and inferring time. |
Tasks | Document Embedding, Sentence Embedding, Sentence Embeddings, Text Summarization |
Published | 2019-07-16 |
URL | https://arxiv.org/abs/1907.07323v1 |
https://arxiv.org/pdf/1907.07323v1.pdf | |
PWC | https://paperswithcode.com/paper/strass-a-light-and-effective-method-for |
Repo | |
Framework | |
High Accurate Unhealthy Leaf Detection
Title | High Accurate Unhealthy Leaf Detection |
Authors | S. Mohan Sai, G. Gopichand, C. Vikas Reddy, K. Mona Teja |
Abstract | India is an agriculture-dependent country. As we all know that farming is the backbone of our country it is our responsibility to preserve the crops. However, we cannot stop the destruction of crops by natural calamities at least we have to try to protect our crops from diseases. To, detect a plant disease we need a fast automatic way. So, this paper presents a model to identify the particular disease of plant leaves at early stages so that we can prevent or take a remedy to stop spreading of the disease. This proposed model is made into five sessions. Image preprocessing includes the enhancement of the low light image done using inception modules in CNN. Low-resolution image enhancement is done using an Adversarial Neural Network. This also includes Conversion of RGB Image to YCrCb color space. Next, this paper presents a methodology for image segmentation which is an important aspect for identifying the disease symptoms. This segmentation is done using the genetic algorithm. Due to this process the segmentation of the leaf Image this helps in detection of the leaf mage automatically and classifying. Texture extraction is done using the statistical model called GLCM and finally, the classification of the diseases is done using the SVM using Different Kernels with the high accuracy. |
Tasks | Image Enhancement, Semantic Segmentation |
Published | 2019-08-14 |
URL | https://arxiv.org/abs/1908.09003v1 |
https://arxiv.org/pdf/1908.09003v1.pdf | |
PWC | https://paperswithcode.com/paper/high-accurate-unhealthy-leaf-detection |
Repo | |
Framework | |
Multi-Zone Unit for Recurrent Neural Networks
Title | Multi-Zone Unit for Recurrent Neural Networks |
Authors | Fandong Meng, Jinchao Zhang, Yang Liu, Jie Zhou |
Abstract | Recurrent neural networks (RNNs) have been widely used to deal with sequence learning problems. The input-dependent transition function, which folds new observations into hidden states to sequentially construct fixed-length representations of arbitrary-length sequences, plays a critical role in RNNs. Based on single space composition, transition functions in existing RNNs often have difficulty in capturing complicated long-range dependencies. In this paper, we introduce a new Multi-zone Unit (MZU) for RNNs. The key idea is to design a transition function that is capable of modeling multiple space composition. The MZU consists of three components: zone generation, zone composition, and zone aggregation. Experimental results on multiple datasets of the character-level language modeling task and the aspect-based sentiment analysis task demonstrate the superiority of the MZU. |
Tasks | Aspect-Based Sentiment Analysis, Language Modelling, Sentiment Analysis |
Published | 2019-11-17 |
URL | https://arxiv.org/abs/1911.07184v1 |
https://arxiv.org/pdf/1911.07184v1.pdf | |
PWC | https://paperswithcode.com/paper/multi-zone-unit-for-recurrent-neural-networks |
Repo | |
Framework | |
Q-GADMM: Quantized Group ADMM for Communication Efficient Decentralized Machine Learning
Title | Q-GADMM: Quantized Group ADMM for Communication Efficient Decentralized Machine Learning |
Authors | Anis Elgabli, Jihong Park, Amrit S. Bedi, Chaouki Ben Issaid, Mehdi Bennis, Vaneet Aggarwal |
Abstract | In this article, we propose a communication-efficient decentralized machine learning (ML) algorithm, coined quantized group ADMM (Q-GADMM). To reduce the number of communication links, every worker in Q-GADMM communicates only with two neighbors, while updating its model via the group alternating direct method of multiplier (GADMM). Next, each worker quantizes its model updates before transmission, thereby decreasing the communication payload size. However, due to the lack of centralized entity in decentralized ML, the communication link sparsification and payload compression may incur error propagation, hindering model training convergence. To overcome this, we develop a novel stochastic quantization method to adaptively adjust model quantization levels and their probabilities, while proving the convergence of Q-GADMM for convex objective functions. Furthermore, to demonstrate the feasibility of Q-GADMM for non-convex objective functions, we propose quantized stochastic GADMM (Q-SGADMM) that incorporates deep neural network architectures and stochastic gradient decent (SGD). Simulation results corroborate that Q-GADMM yields 7x less total communication cost while achieving almost the same accuracy and convergence speed compared to GADMM without quantization for a linear regression task. Similarly, for an image classification task, Q-SGADMM achieves 4x less total communication cost with identical accuracy and convergence speed compared to its counterpart without quantization, i.e., stochastic GADMM (SGADMM). |
Tasks | Image Classification, Quantization |
Published | 2019-10-23 |
URL | https://arxiv.org/abs/1910.10453v3 |
https://arxiv.org/pdf/1910.10453v3.pdf | |
PWC | https://paperswithcode.com/paper/q-gadmm-quantized-group-admm-for |
Repo | |
Framework | |
f-VAEGAN-D2: A Feature Generating Framework for Any-Shot Learning
Title | f-VAEGAN-D2: A Feature Generating Framework for Any-Shot Learning |
Authors | Yongqin Xian, Saurabh Sharma, Bernt Schiele, Zeynep Akata |
Abstract | When labeled training data is scarce, a promising data augmentation approach is to generate visual features of unknown classes using their attributes. To learn the class conditional distribution of CNN features, these models rely on pairs of image features and class attributes. Hence, they can not make use of the abundance of unlabeled data samples. In this paper, we tackle any-shot learning problems i.e. zero-shot and few-shot, in a unified feature generating framework that operates in both inductive and transductive learning settings. We develop a conditional generative model that combines the strength of VAE and GANs and in addition, via an unconditional discriminator, learns the marginal feature distribution of unlabeled images. We empirically show that our model learns highly discriminative CNN features for five datasets, i.e. CUB, SUN, AWA and ImageNet, and establish a new state-of-the-art in any-shot learning, i.e. inductive and transductive (generalized) zero- and few-shot learning settings. We also demonstrate that our learned features are interpretable: we visualize them by inverting them back to the pixel space and we explain them by generating textual arguments of why they are associated with a certain label. |
Tasks | Data Augmentation, Few-Shot Learning |
Published | 2019-03-25 |
URL | http://arxiv.org/abs/1903.10132v1 |
http://arxiv.org/pdf/1903.10132v1.pdf | |
PWC | https://paperswithcode.com/paper/f-vaegan-d2-a-feature-generating-framework |
Repo | |
Framework | |
The Sparse Reverse of Principal Component Analysis for Fast Low-Rank Matrix Completion
Title | The Sparse Reverse of Principal Component Analysis for Fast Low-Rank Matrix Completion |
Authors | Abdallah Chehade, Zunya Shi |
Abstract | Matrix completion constantly receives tremendous attention from many research fields. It is commonly applied for recommender systems such as movie ratings, computer vision such as image reconstruction or completion, multi-task learning such as collaboratively modeling time-series trends of multiple sensors, and many other applications. Matrix completion techniques are usually computationally exhaustive and/or fail to capture the heterogeneity in the data. For example, images usually contain a heterogeneous set of objects, and thus it is a challenging task to reconstruct images with high levels of missing data. In this paper, we propose the sparse reverse of principal component analysis for matrix completion. The proposed approach maintains smoothness across the matrix, produces accurate estimates of the missing data, converges iteratively, and it is computationally tractable with a controllable upper bound on the number of iterations until convergence. The accuracy of the proposed technique is validated on natural images, movie ratings, and multisensor data. It is also compared with common benchmark methods used for matrix completion. |
Tasks | Image Reconstruction, Low-Rank Matrix Completion, Matrix Completion, Multi-Task Learning, Recommendation Systems, Time Series |
Published | 2019-10-04 |
URL | https://arxiv.org/abs/1910.02155v1 |
https://arxiv.org/pdf/1910.02155v1.pdf | |
PWC | https://paperswithcode.com/paper/the-sparse-reverse-of-principal-component |
Repo | |
Framework | |
Jointly Adversarial Network to Wavelength Compensation and Dehazing of Underwater Images
Title | Jointly Adversarial Network to Wavelength Compensation and Dehazing of Underwater Images |
Authors | Xueyan Ding, Yafei Wang, Yang Yan, Zheng Liang, Zetian Mi, Xianping Fu |
Abstract | Severe color casts, low contrast and blurriness of underwater images caused by light absorption and scattering result in a difficult task for exploring underwater environments. Different from most of previous underwater image enhancement methods that compute light attenuation along object-camera path through hazy image formation model, we propose a novel jointly wavelength compensation and dehazing network (JWCDN) that takes into account the wavelength attenuation along surface-object path and the scattering along object-camera path simultaneously. By embedding a simplified underwater formation model into generative adversarial network, we can jointly estimates the transmission map, wavelength attenuation and background light via different network modules, and uses the simplified underwater image formation model to recover degraded underwater images. Especially, a multi-scale densely connected encoder-decoder network is proposed to leverage features from multiple layers for estimating the transmission map. To further improve the recovered image, we use an edge preserving network module to enhance the detail of the recovered image. Moreover, to train the proposed network, we propose a novel underwater image synthesis method that generates underwater images with inherent optical properties of different water types. The synthesis method can simulate the color, contrast and blurriness appearance of real-world underwater environments simultaneously. Extensive experiments on synthetic and real-world underwater images demonstrate that the proposed method yields comparable or better results on both subjective and objective assessments, compared with several state-of-the-art methods. |
Tasks | Image Enhancement, Image Generation |
Published | 2019-07-12 |
URL | https://arxiv.org/abs/1907.05595v1 |
https://arxiv.org/pdf/1907.05595v1.pdf | |
PWC | https://paperswithcode.com/paper/jointly-adversarial-network-to-wavelength |
Repo | |
Framework | |
Anisotropic Super Resolution in Prostate MRI using Super Resolution Generative Adversarial Networks
Title | Anisotropic Super Resolution in Prostate MRI using Super Resolution Generative Adversarial Networks |
Authors | Rewa Sood, Mirabela Rusu |
Abstract | Acquiring High Resolution (HR) Magnetic Resonance (MR) images requires the patient to remain still for long periods of time, which causes patient discomfort and increases the probability of motion induced image artifacts. A possible solution is to acquire low resolution (LR) images and to process them with the Super Resolution Generative Adversarial Network (SRGAN) to create a super-resolved version. This work applies SRGAN to MR images of the prostate and performs three experiments. The first experiment explores improving the in-plane MR image resolution by factors of 4 and 8, and shows that, while the PSNR and SSIM (Structural SIMilarity) metrics are lower than the isotropic bicubic interpolation baseline, the SRGAN is able to create images that have high edge fidelity. The second experiment explores anisotropic super-resolution via synthetic images, in that the input images to the network are anisotropically downsampled versions of HR images. This experiment demonstrates the ability of the modified SRGAN to perform anisotropic super-resolution, with quantitative image metrics that are comparable to those of the anisotropic bicubic interpolation baseline. Finally, the third experiment applies a modified version of the SRGAN to super-resolve anisotropic images obtained from the through-plane slices of the volumetric MR data. The output super-resolved images contain a significant amount of high frequency information that make them visually close to their HR counterparts. Overall, the promising results from each experiment show that super-resolution for MR images is a successful technique and that producing isotropic MR image volumes from anisotropic slices is an achievable goal. |
Tasks | Super-Resolution |
Published | 2019-12-19 |
URL | https://arxiv.org/abs/1912.09497v1 |
https://arxiv.org/pdf/1912.09497v1.pdf | |
PWC | https://paperswithcode.com/paper/anisotropic-super-resolution-in-prostate-mri |
Repo | |
Framework | |
Ordinal Bucketing for Game Trees using Dynamic Quantile Approximation
Title | Ordinal Bucketing for Game Trees using Dynamic Quantile Approximation |
Authors | Tobias Joppen, Tilman Strübig, Johannes Fürnkranz |
Abstract | In this paper, we present a simple and cheap ordinal bucketing algorithm that approximately generates $q$-quantiles from an incremental data stream. The bucketing is done dynamically in the sense that the amount of buckets $q$ increases with the number of seen samples. We show how this can be used in Ordinal Monte Carlo Tree Search (OMCTS) to yield better bounds on time and space complexity, especially in the presence of noisy rewards. Besides complexity analysis and quality tests of quantiles, we evaluate our method using OMCTS in the General Video Game Framework (GVGAI). Our results demonstrate its dominance over vanilla Monte Carlo Tree Search in the presence of noise, where OMCTS without bucketing has a very bad time and space complexity. |
Tasks | |
Published | 2019-05-31 |
URL | https://arxiv.org/abs/1905.13449v1 |
https://arxiv.org/pdf/1905.13449v1.pdf | |
PWC | https://paperswithcode.com/paper/ordinal-bucketing-for-game-trees-using |
Repo | |
Framework | |
Targeted Estimation of Heterogeneous Treatment Effect in Observational Survival Analysis
Title | Targeted Estimation of Heterogeneous Treatment Effect in Observational Survival Analysis |
Authors | Jie Zhu, Blanca Gallego |
Abstract | The aim of clinical effectiveness research using repositories of electronic health records is to identify what health interventions ‘work best’ in real-world settings. Since there are several reasons why the net benefit of intervention may differ across patients, current comparative effectiveness literature focuses on investigating heterogeneous treatment effect and predicting whether an individual might benefit from an intervention. The majority of this literature has concentrated on the estimation of the effect of treatment on binary outcomes. However, many medical interventions are evaluated in terms of their effect on future events, which are subject to loss to follow-up. In this study, we describe a framework for the estimation of heterogeneous treatment effect in terms of differences in time-to-event (survival) probabilities. We divide the problem into three phases: (1) estimation of treatment effect conditioned on unique sets of the covariate vector; (2) identification of features important for heterogeneity using an ensemble of non-parametric variable importance methods; and (3) estimation of treatment effect on the reference classes defined by the previously selected features, using one-step Targeted Maximum Likelihood Estimation. We conducted a series of simulation studies and found that this method performs well when either sample size or event rate is high enough and the number of covariates contributing to the effect heterogeneity is moderate. An application of this method to a clinical case study was conducted by estimating the effect of oral anticoagulants on newly diagnosed non-valvular atrial fibrillation patients using data from the UK Clinical Practice Research Datalink. |
Tasks | Survival Analysis |
Published | 2019-10-20 |
URL | https://arxiv.org/abs/1910.08877v2 |
https://arxiv.org/pdf/1910.08877v2.pdf | |
PWC | https://paperswithcode.com/paper/targeted-estimation-of-heterogeneous |
Repo | |
Framework | |