Paper Group ANR 738
Beyond Photo Realism for Domain Adaptation from Synthetic Data. Fast, Small, and Simple Document Listing on Repetitive Text Collections. Online Regularization by Denoising with Applications to Phase Retrieval. A Streaming Analytics Language for Processing Cyber Data. LIA: Latently Invertible Autoencoder with Adversarial Learning. On the Anomalous G …
Beyond Photo Realism for Domain Adaptation from Synthetic Data
Title | Beyond Photo Realism for Domain Adaptation from Synthetic Data |
Authors | Kristofer Schlachter, Connor DeFanti, Sebastian Herscher, Ken Perlin, Jonathan Tompson |
Abstract | As synthetic imagery is used more frequently in training deep models, it is important to understand how different synthesis techniques impact the performance of such models. In this work, we perform a thorough evaluation of the effectiveness of several different synthesis techniques and their impact on the complexity of classifier domain adaptation to the “real” underlying data distribution that they seek to replicate. In addition, we propose a novel learned synthesis technique to better train classifier models than state-of-the-art offline graphical methods, while using significantly less computational resources. We accomplish this by learning a generative model to perform shading of synthetic geometry conditioned on a “g-buffer” representation of the scene to render, as well as a low sample Monte Carlo rendered image. The major contributions are (i) a dataset that allows comparison of real and synthetic versions of the same scene, (ii) an augmented data representation that boosts the stability of learning and improves the datasets accuracy, (iii) three different partially differentiable rendering techniques where lighting, denoising and shading are learned, and (iv) we improve a state of the art generative adversarial network (GAN) approach by using an ensemble of trained models to generate datasets that approach the performance of training on real data and surpass the performance of the full global illumination rendering. |
Tasks | Denoising, Domain Adaptation |
Published | 2019-09-04 |
URL | https://arxiv.org/abs/1909.01960v1 |
https://arxiv.org/pdf/1909.01960v1.pdf | |
PWC | https://paperswithcode.com/paper/beyond-photo-realism-for-domain-adaptation |
Repo | |
Framework | |
Fast, Small, and Simple Document Listing on Repetitive Text Collections
Title | Fast, Small, and Simple Document Listing on Repetitive Text Collections |
Authors | Dustin Cobas, Gonzalo Navarro |
Abstract | Document listing on string collections is the task of finding all documents where a pattern appears. It is regarded as the most fundamental document retrieval problem, and is useful in various applications. Many of the fastest-growing string collections are composed of very similar documents, such as versioned code and document collections, genome repositories, etc. Plain pattern-matching indexes designed for repetitive text collections achieve orders-of-magnitude reductions in space. Instead, there are not many analogous indexes for document retrieval. In this paper we present a simple document listing index for repetitive string collections of total length $n$ that lists the $ndoc$ distinct documents where a pattern of length $m$ appears in time $\mathcal{O}(m+ndoc \cdot \log n)$. We exploit the repetitiveness of the document array (i.e., the suffix array coarsened to document identifiers) to grammar-compress it while precomputing the answers to nonterminals, and store them in grammar-compressed form as well. Our experimental results show that our index sharply outperforms existing alternatives in the space/time tradeoff map. |
Tasks | |
Published | 2019-02-20 |
URL | http://arxiv.org/abs/1902.07599v1 |
http://arxiv.org/pdf/1902.07599v1.pdf | |
PWC | https://paperswithcode.com/paper/fast-small-and-simple-document-listing-on |
Repo | |
Framework | |
Online Regularization by Denoising with Applications to Phase Retrieval
Title | Online Regularization by Denoising with Applications to Phase Retrieval |
Authors | Zihui Wu, Yu Sun, Jiaming Liu, Ulugbek S. Kamilov |
Abstract | Regularization by denoising (RED) is a powerful framework for solving imaging inverse problems. Most RED algorithms are iterative batch procedures, which limits their applicability to very large datasets. In this paper, we address this limitation by introducing a novel online RED (On-RED) algorithm, which processes a small subset of the data at a time. We establish the theoretical convergence of On-RED in convex settings and empirically discuss its effectiveness in non-convex ones by illustrating its applicability to phase retrieval. Our results suggest that On-RED is an effective alternative to the traditional RED algorithms when dealing with large datasets. |
Tasks | Denoising |
Published | 2019-09-04 |
URL | https://arxiv.org/abs/1909.02040v1 |
https://arxiv.org/pdf/1909.02040v1.pdf | |
PWC | https://paperswithcode.com/paper/online-regularization-by-denoising-with |
Repo | |
Framework | |
A Streaming Analytics Language for Processing Cyber Data
Title | A Streaming Analytics Language for Processing Cyber Data |
Authors | Eric L. Goodman, Dirk Grunwald |
Abstract | We present a domain-specific language called SAL(the Streaming Analytics Language) for processing data in a semi-streaming model. In particular we examine the use case of processing netflow data in order to identify malicious actors within a network. Because of the large volume of data generated from networks, it is often only feasible to process the data with a single pass, utilizing a streaming (O(polylog n) space requirements) or semi-streaming computing model ( O(n polylog n) space requirements). Despite these constraints, we are able to achieve an average of 0.87 for the AUC of the ROC curve for a set of situations dealing with botnet detection. The implementation of an interpreter for SAL, which we call SAM (Streaming Analytics Machine), achieves scaling results that show improved throughput to 61 nodes (976 cores), with an overall rate of 373,000 netflows per second or 32.2 billion per day. SAL provides a succinct way to describe common analyses that allow cyber analysts to find data of interest, and SAM is a scalable interpreter of the language. |
Tasks | |
Published | 2019-11-03 |
URL | https://arxiv.org/abs/1911.00815v1 |
https://arxiv.org/pdf/1911.00815v1.pdf | |
PWC | https://paperswithcode.com/paper/a-streaming-analytics-language-for-processing |
Repo | |
Framework | |
LIA: Latently Invertible Autoencoder with Adversarial Learning
Title | LIA: Latently Invertible Autoencoder with Adversarial Learning |
Authors | Jiapeng Zhu, Deli Zhao, Bolei Zhou, Bo Zhang |
Abstract | Deep generative models such as Variational AutoEncoder (VAE) and Generative Adversarial Network (GAN) play an increasingly important role in machine learning and computer vision. However, there are two fundamental issues hindering their real-world applications: the difficulty of conducting variational inference in VAE and the functional absence of encoding real-world samples in GAN. In this paper, we propose a novel algorithm named Latently Invertible Autoencoder (LIA) to address the above two issues in one framework. An invertible network and its inverse mapping are symmetrically embedded in the latent space of VAE. Thus the partial encoder first transforms the input into feature vectors and then the distribution of these feature vectors is reshaped to fit a prior by the invertible network. The decoder proceeds in the reverse order of the encoder’s composite mappings. A two-stage stochasticity-free training scheme is designed to train LIA via adversarial learning, in the sense that the decoder of LIA is first trained as a standard GAN with the invertible network and then the partial encoder is learned from an autoencoder by detaching the invertible network from LIA. Experiments conducted on the FFHQ face dataset and three LSUN datasets validate the effectiveness of LIA for inference and generation. |
Tasks | |
Published | 2019-06-19 |
URL | https://arxiv.org/abs/1906.08090v2 |
https://arxiv.org/pdf/1906.08090v2.pdf | |
PWC | https://paperswithcode.com/paper/lia-latently-invertible-autoencoder-with |
Repo | |
Framework | |
On the Anomalous Generalization of GANs
Title | On the Anomalous Generalization of GANs |
Authors | Jinchen Xuan, Yunchang Yang, Ze Yang, Di He, Liwei Wang |
Abstract | Generative models, especially Generative Adversarial Networks (GANs), have received significant attention recently. However, it has been observed that in terms of some attributes, e.g. the number of simple geometric primitives in an image, GANs are not able to learn the target distribution in practice. Motivated by this observation, we discover two specific problems of GANs leading to anomalous generalization behaviour, which we refer to as the sample insufficiency and the pixel-wise combination. For the first problem of sample insufficiency, we show theoretically and empirically that the batchsize of the training samples in practice may be insufficient for the discriminator to learn an accurate discrimination function. It could result in unstable training dynamics for the generator, leading to anomalous generalization. For the second problem of pixel-wise combination, we find that besides recognizing the positive training samples as real, under certain circumstances, the discriminator could be fooled to recognize the pixel-wise combinations (e.g. pixel-wise average) of the positive training samples as real. However, those combinations could be visually different from the real samples in the target distribution. With the fooled discriminator as reference, the generator would obtain biased supervision further, leading to the anomalous generalization behaviour. Additionally, in this paper, we propose methods to mitigate the anomalous generalization of GANs. Extensive experiments on benchmark show our proposed methods improve the FID score up to 30% on natural image dataset. |
Tasks | |
Published | 2019-09-27 |
URL | https://arxiv.org/abs/1909.12638v2 |
https://arxiv.org/pdf/1909.12638v2.pdf | |
PWC | https://paperswithcode.com/paper/on-the-anomalous-generalization-of-gans-1 |
Repo | |
Framework | |
EnSyth: A Pruning Approach to Synthesis of Deep Learning Ensembles
Title | EnSyth: A Pruning Approach to Synthesis of Deep Learning Ensembles |
Authors | Besher Alhalabi, Mohamed Medhat Gaber, Shadi Basurra |
Abstract | Deep neural networks have achieved state-of-art performance in many domains including computer vision, natural language processing and self-driving cars. However, they are very computationally expensive and memory intensive which raises significant challenges when it comes to deploy or train them on strict latency applications or resource-limited environments. As a result, many attempts have been introduced to accelerate and compress deep learning models, however the majority were not able to maintain the same accuracy of the baseline models. In this paper, we describe EnSyth, a deep learning ensemble approach to enhance the predictability of compact neural network’s models. First, we generate a set of diverse compressed deep learning models using different hyperparameters for a pruning method, after that we utilise ensemble learning to synthesise the outputs of the compressed models to compose a new pool of classifiers. Finally, we apply backward elimination on the generated pool to explore the best performing combinations of models. On CIFAR-10, CIFAR-5 data-sets with LeNet-5, EnSyth outperforms the predictability of the baseline model. |
Tasks | Self-Driving Cars |
Published | 2019-07-22 |
URL | https://arxiv.org/abs/1907.09286v1 |
https://arxiv.org/pdf/1907.09286v1.pdf | |
PWC | https://paperswithcode.com/paper/ensyth-a-pruning-approach-to-synthesis-of |
Repo | |
Framework | |
Bipolar in Temporal Argumentation Framework
Title | Bipolar in Temporal Argumentation Framework |
Authors | Maximiliano C. D. Budán, Maria Laura Cobo, Diego C. Martinez, Guillermo R. Simari |
Abstract | A Timed Argumentation Framework (TAF) is a formalism where arguments are only valid for consideration in a given period of time, called availability intervals, which are defined for every individual argument. The original proposal is based on a single, abstract notion of attack between arguments that remains static and permanent in time. Thus, in general, when identifying the set of acceptable arguments, the outcome associated with a TAF will vary over time. In this work we introduce an extension of TAF adding the capability of modeling a support relation between arguments. In this sense, the resulting framework provides a suitable model for different time-dependent issues. Thus, the main contribution here is to provide an enhanced framework for modeling a positive (support) and negative (attack) interaction varying over time, which are relevant in many real-world situations. This leads to a Timed Bipolar Argumentation Framework (T-BAF), where classical argument extensions can be defined. The proposal aims at advancing in the integration of temporal argumentation in different application domain. |
Tasks | |
Published | 2019-03-05 |
URL | http://arxiv.org/abs/1903.01874v1 |
http://arxiv.org/pdf/1903.01874v1.pdf | |
PWC | https://paperswithcode.com/paper/bipolar-in-temporal-argumentation-framework |
Repo | |
Framework | |
Dynamic Cell Imaging in PET with Optimal Transport Regularization
Title | Dynamic Cell Imaging in PET with Optimal Transport Regularization |
Authors | Bernhard Schmitzer, Klaus P. Schäfers, Benedikt Wirth |
Abstract | We propose a novel dynamic image reconstruction method from PET listmode data that could be particularly suited to tracking single or small numbers of cells. In contrast to conventional PET reconstruction our method combines the information from all detected events not only to reconstruct the dynamic evolution of the radionuclide distribution, but also to improve the reconstruction at each single time point by enforcing temporal consistency. This is achieved via optimal transport regularization where in principle, among all possible temporally evolving radionuclide distributions consistent with the PET measurement, the one is chosen with least kinetic motion energy. The reconstruction is found by convex optimization so that there is no dependence on the initialization of the method. We study its behaviour on simulated data of a human PET system and demonstrate its robustness even in settings with very low radioactivity. In contrast to previously reported cell tracking algorithms, our technique is oblivious to the number of tracked cells. Without any additional complexity one or multiple cells can be reconstructed, and the model automatically determines the number of particles. For instance, four radiolabelled cells moving at a velocity of 3.1 mm/s and a PET recorded count rate of 1.1 cps (for each cell) could be simultaneously tracked with a tracking accuracy of 5.3 mm inside a simulated human body. |
Tasks | Image Reconstruction |
Published | 2019-02-20 |
URL | https://arxiv.org/abs/1902.07521v2 |
https://arxiv.org/pdf/1902.07521v2.pdf | |
PWC | https://paperswithcode.com/paper/dynamic-cell-imaging-in-pet-with-optimal |
Repo | |
Framework | |
Smart Predict-and-Optimize for Hard Combinatorial Optimization Problems
Title | Smart Predict-and-Optimize for Hard Combinatorial Optimization Problems |
Authors | Jaynta Mandi, Emir Demirović, Peter. J Stuckey, Tias Guns |
Abstract | Combinatorial optimization assumes that all parameters of the optimization problem, e.g. the weights in the objective function is fixed. Often, these weights are mere estimates and increasingly machine learning techniques are used to for their estimation. Recently, Smart Predict and Optimize (SPO) has been proposed for problems with a linear objective function over the predictions, more specifically linear programming problems. It takes the regret of the predictions on the linear problem into account, by repeatedly solving it during learning. We investigate the use of SPO to solve more realistic discrete optimization problems. The main challenge is the repeated solving of the optimization problem. To this end, we investigate ways to relax the problem as well as warmstarting the learning and the solving. Our results show that even for discrete problems it often suffices to train by solving the relaxation in the SPO loss. Furthermore, this approach outperforms, for most instances, the state-of-the-art approach of Wilder, Dilkina, and Tambe. We experiment with weighted knapsack problems as well as complex scheduling problems and show for the first time that a predict-and-optimize approach can successfully be used on large-scale combinatorial optimization problems. |
Tasks | Combinatorial Optimization |
Published | 2019-11-22 |
URL | https://arxiv.org/abs/1911.10092v1 |
https://arxiv.org/pdf/1911.10092v1.pdf | |
PWC | https://paperswithcode.com/paper/smart-predict-and-optimize-for-hard |
Repo | |
Framework | |
Intelligent Processing in Vehicular Ad hoc Networks: a Survey
Title | Intelligent Processing in Vehicular Ad hoc Networks: a Survey |
Authors | Yang Liu |
Abstract | The intelligent Processing technique is more and more attractive to researchers due to its ability to deal with key problems in Vehicular Ad hoc networks. However, several problems in applying intelligent processing technologies in VANETs remain open. The existing applications are comprehensively reviewed and discussed, and classified into different categories in this paper. Their strategies, advantages/disadvantages, and performances are elaborated. By generalizing different tactics in various applications related to different scenarios of VANETs and evaluating their performances, several promising directions for future research have been suggested. |
Tasks | |
Published | 2019-03-28 |
URL | http://arxiv.org/abs/1903.11916v1 |
http://arxiv.org/pdf/1903.11916v1.pdf | |
PWC | https://paperswithcode.com/paper/intelligent-processing-in-vehicular-ad-hoc |
Repo | |
Framework | |
Weakly Supervised Deep Learning Approach in Streaming Environments
Title | Weakly Supervised Deep Learning Approach in Streaming Environments |
Authors | Mahardhika Pratama, Andri Ashfahani, Mohamad Abdul Hady |
Abstract | The feasibility of existing data stream algorithms is often hindered by the weakly supervised condition of data streams. A self-evolving deep neural network, namely Parsimonious Network (ParsNet), is proposed as a solution to various weakly-supervised data stream problems. A self-labelling strategy with hedge (SLASH) is proposed in which its auto-correction mechanism copes with \textit{the accumulation of mistakes} significantly affecting the model’s generalization. ParsNet is developed from a closed-loop configuration of the self-evolving generative and discriminative training processes exploiting shared parameters in which its structure flexibly grows and shrinks to overcome the issue of concept drift with/without labels. The numerical evaluation has been performed under two challenging problems, namely sporadic access to ground truth and infinitely delayed access to the ground truth. Our numerical study shows the advantage of ParsNet with a substantial margin from its counterparts in the high-dimensional data streams and infinite delay simulation protocol. To support the reproducible research initiative, the source code of ParsNet along with supplementary materials are made available at https://bit.ly/2qNW7p4. |
Tasks | |
Published | 2019-11-03 |
URL | https://arxiv.org/abs/1911.00847v3 |
https://arxiv.org/pdf/1911.00847v3.pdf | |
PWC | https://paperswithcode.com/paper/weakly-supervised-deep-learning-approach-in |
Repo | |
Framework | |
An End-to-End HW/SW Co-Design Methodology to Design Efficient Deep Neural Network Systems using Virtual Models
Title | An End-to-End HW/SW Co-Design Methodology to Design Efficient Deep Neural Network Systems using Virtual Models |
Authors | Michael J. Klaiber, Sebastian Vogel, Axel Acosta, Robert Korn, Leonardo Ecco, Kristine Back, Andre Guntoro, Ingo Feldner |
Abstract | End-to-end performance estimation and measurement of deep neural network (DNN) systems become more important with increasing complexity of DNN systems consisting of hardware and software components. The methodology proposed in this paper aims at a reduced turn-around time for evaluating different design choices of hardware and software components of DNN systems. This reduction is achieved by moving the performance estimation from the implementation phase to the concept phase by employing virtual hardware models instead of gathering measurement results from physical prototypes. Deep learning compilers introduce hardware-specific transformations and are, therefore, considered a part of the design flow of virtual system models to extract end-to-end performance estimations. To validate the run-time accuracy of the proposed methodology, a system processing the DilatedVGG DNN is realized both as virtual system model and as hardware implementation. The results show that up to 92 % accuracy can be reached in predicting the processing time of the DNN inference. |
Tasks | |
Published | 2019-10-25 |
URL | https://arxiv.org/abs/1910.11632v2 |
https://arxiv.org/pdf/1910.11632v2.pdf | |
PWC | https://paperswithcode.com/paper/an-end-to-end-hwsw-co-design-methodology-to |
Repo | |
Framework | |
Feature Engineering Combined with 1 D Convolutional Neural Network for Improved Mortality Prediction
Title | Feature Engineering Combined with 1 D Convolutional Neural Network for Improved Mortality Prediction |
Authors | Saumil Maheshwari, Rohit Verma, Anupam Shukla, Ritu Tiwari, Rishu Garg |
Abstract | The intensive care units (ICUs) are responsible for generating a wealth of useful data in the form of Electronic Health Record (EHR). This data allows for the development of a prediction tool with perfect knowledge backing. We aimed to build a mortality prediction model on 2012 Physionet Challenge mortality prediction database of 4000 patients admitted in ICU. The challenges in the dataset, such as high dimensionality, imbalanced distribution, and missing values were tackled with analytical methods and tools via feature engineering and new variable construction. The objective of the research is to utilize the relations among the clinical variables and construct new variables which would establish the effectiveness of 1-Dimensional Convolutional Neural Network (1- D CNN) with constructed features. Its performance with the traditional machine learning algorithms like XGBoost classifier, Support Vector Machine (SVM), K-Neighbours Classifier (K-NN), and Random Forest Classifier (RF) is compared for Area Under Curve (AUC). The investigation reveals the best AUC of 0.848 using 1-D CNN model. |
Tasks | Feature Engineering, Mortality Prediction |
Published | 2019-12-08 |
URL | https://arxiv.org/abs/1912.03789v1 |
https://arxiv.org/pdf/1912.03789v1.pdf | |
PWC | https://paperswithcode.com/paper/feature-engineering-combined-with-1-d |
Repo | |
Framework | |
A Hybrid Retrieval-Generation Neural Conversation Model
Title | A Hybrid Retrieval-Generation Neural Conversation Model |
Authors | Liu Yang, Junjie Hu, Minghui Qiu, Chen Qu, Jianfeng Gao, W. Bruce Croft, Xiaodong Liu, Yelong Shen, Jingjing Liu |
Abstract | Intelligent personal assistant systems that are able to have multi-turn conversations with human users are becoming increasingly popular. Most previous research has been focused on using either retrieval-based or generation-based methods to develop such systems. Retrieval-based methods have the advantage of returning fluent and informative responses with great diversity. However, the performance of the methods is limited by the size of the response repository. On the other hand, generation-based methods can produce highly coherent responses on any topics. But the generated responses are often generic and not informative due to the lack of grounding knowledge. In this paper, we propose a hybrid neural conversation model that combines the merits of both response retrieval and generation methods. Experimental results on Twitter and Foursquare data show that the proposed model outperforms both retrieval-based methods and generation-based methods (including a recently proposed knowledge-grounded neural conversation model) under both automatic evaluation metrics and human evaluation. We hope that the findings in this study provide new insights on how to integrate text retrieval and text generation models for building conversation systems. |
Tasks | Text Generation |
Published | 2019-04-19 |
URL | https://arxiv.org/abs/1904.09068v2 |
https://arxiv.org/pdf/1904.09068v2.pdf | |
PWC | https://paperswithcode.com/paper/a-hybrid-retrieval-generation-neural |
Repo | |
Framework | |