January 28, 2020

3092 words 15 mins read

Paper Group ANR 962

Paper Group ANR 962

Heterogeneous Collaborative Filtering. Edge AI: On-Demand Accelerating Deep Neural Network Inference via Edge Computing. Manifold Forests: Closing the Gap on Neural Networks. Finding Interpretable Concept Spaces in Node Embeddings using Knowledge Bases. Deep Learning and Gaussian Process based Band Assignment in Dual Band Systems. One-Shot Template …

Heterogeneous Collaborative Filtering

Title Heterogeneous Collaborative Filtering
Authors Yifang Liu, Zhentao Xu, Cong Hui, Yi Xuan, Jessie Chen, Yuanming Shan
Abstract Recommendation system is important to a content sharing/creating social network. Collaborative filtering is a widely-adopted technology in conventional recommenders, which is based on similarity between positively engaged content items involving the same users. Conventional collaborative filtering (CCF) suffers from cold start problem and narrow content diversity. We propose a new recommendation approach, heterogeneous collaborative filtering (HCF) to tackle these challenges at the root, while keeping the strength of collaborative filtering. We present two implementation algorithms of HCF for content recommendation and content dissemination. Experiment results demonstrate that our approach improve the recommendation quality in a real world social network for content creating and sharing.
Tasks
Published 2019-08-31
URL https://arxiv.org/abs/1909.01727v1
PDF https://arxiv.org/pdf/1909.01727v1.pdf
PWC https://paperswithcode.com/paper/heterogeneous-collaborative-filtering
Repo
Framework

Edge AI: On-Demand Accelerating Deep Neural Network Inference via Edge Computing

Title Edge AI: On-Demand Accelerating Deep Neural Network Inference via Edge Computing
Authors En Li, Liekang Zeng, Zhi Zhou, Xu Chen
Abstract As a key technology of enabling Artificial Intelligence (AI) applications in 5G era, Deep Neural Networks (DNNs) have quickly attracted widespread attention. However, it is challenging to run computation-intensive DNN-based tasks on mobile devices due to the limited computation resources. What’s worse, traditional cloud-assisted DNN inference is heavily hindered by the significant wide-area network latency, leading to poor real-time performance as well as low quality of user experience. To address these challenges, in this paper, we propose Edgent, a framework that leverages edge computing for DNN collaborative inference through device-edge synergy. Edgent exploits two design knobs: (1) DNN partitioning that adaptively partitions computation between device and edge for purpose of coordinating the powerful cloud resource and the proximal edge resource for real-time DNN inference; (2) DNN right-sizing that further reduces computing latency via early exiting inference at an appropriate intermediate DNN layer. In addition, considering the potential network fluctuation in real-world deployment, Edgentis properly design to specialize for both static and dynamic network environment. Specifically, in a static environment where the bandwidth changes slowly, Edgent derives the best configurations with the assist of regression-based prediction models, while in a dynamic environment where the bandwidth varies dramatically, Edgent generates the best execution plan through the online change point detection algorithm that maps the current bandwidth state to the optimal configuration. We implement Edgent prototype based on the Raspberry Pi and the desktop PC and the extensive experimental evaluations demonstrate Edgent’s effectiveness in enabling on-demand low-latency edge intelligence.
Tasks Change Point Detection
Published 2019-10-04
URL https://arxiv.org/abs/1910.05316v1
PDF https://arxiv.org/pdf/1910.05316v1.pdf
PWC https://paperswithcode.com/paper/edge-ai-on-demand-accelerating-deep-neural
Repo
Framework

Manifold Forests: Closing the Gap on Neural Networks

Title Manifold Forests: Closing the Gap on Neural Networks
Authors Ronan Perry, Tyler M. Tomita, Jesse Patsolic, Benjamin Falk, Joshua T. Vogelstein
Abstract Decision forests (DF), in particular random forests and gradient boosting trees, have demonstrated state-of-the-art accuracy compared to other methods in many supervised learning scenarios. In particular, DFs dominate other methods in tabular data, that is, when the feature space is unstructured, so that the signal is invariant to permuting feature indices. However, in structured data lying on a manifold—such as images, text, and speech—neural nets (NN) tend to outperform DFs. We conjecture that at least part of the reason for this is that the input to NN is not simply the feature magnitudes, but also their indices (for example, the convolution operation uses “feature locality”). In contrast, na"ive DF implementations fail to explicitly consider feature indices. A recently proposed DF approach demonstrates that DFs, for each node, implicitly sample a random matrix from some specific distribution. Here, we build on that to show that one can choose distributions in a \emph{manifold aware fashion}. For example, for image classification, rather than randomly selecting pixels, one can randomly select contiguous patches. We demonstrate the empirical performance of data living on three different manifolds: images, time-series, and a torus. In all three cases, our Manifold Forest (\Mf) algorithm empirically dominates other state-of-the-art approaches that ignore feature space structure, achieving a lower classification error on all sample sizes. This dominance extends to the MNIST data set as well. Moreover, both training and test time is significantly faster for manifold forests as compared to deep nets. This approach, therefore, has promise to enable DFs and other machine learning methods to close the gap with deep nets on manifold-valued data.
Tasks Image Classification, Time Series
Published 2019-09-25
URL https://arxiv.org/abs/1909.11799v1
PDF https://arxiv.org/pdf/1909.11799v1.pdf
PWC https://paperswithcode.com/paper/manifold-forests-closing-the-gap-on-neural
Repo
Framework

Finding Interpretable Concept Spaces in Node Embeddings using Knowledge Bases

Title Finding Interpretable Concept Spaces in Node Embeddings using Knowledge Bases
Authors Maximilian Idahl, Megha Khosla, Avishek Anand
Abstract In this paper we propose and study the novel problem of explaining node embeddings by finding embedded human interpretable subspaces in already trained unsupervised node representation embeddings. We use an external knowledge base that is organized as a taxonomy of human-understandable concepts over entities as a guide to identify subspaces in node embeddings learned from an entity graph derived from Wikipedia. We propose a method that given a concept finds a linear transformation to a subspace where the structure of the concept is retained. Our initial experiments show that we obtain low error in finding fine-grained concepts.
Tasks
Published 2019-10-11
URL https://arxiv.org/abs/1910.05030v1
PDF https://arxiv.org/pdf/1910.05030v1.pdf
PWC https://paperswithcode.com/paper/finding-interpretable-concept-spaces-in-node
Repo
Framework

Deep Learning and Gaussian Process based Band Assignment in Dual Band Systems

Title Deep Learning and Gaussian Process based Band Assignment in Dual Band Systems
Authors Daoud Burghal, Rui Wang, Andreas F. Molisch
Abstract We consider the band assignment (BA) problem in dual-band systems, where the basestation (BS) chooses one of the two available frequency bands (centimeter-wave and millimeter-wave bands) to communicate with the user equipment (UE). While the millimeter-wave band might offer higher data rate, there is a significant probability of outage during which the communication should be carried on the (more reliable) centimeter-wave band. We consider two variations of the BA problem, one-shot and sequential BA. For the former the BS uses only the currently observed information to decide whether to switch to the other frequency band, for the sequential BA, the BS uses a window of previously observed information to predict the best band for a future time step. We provide two approaches to solve the BA problem, (i) a deep learning approach that is based on Long Short Term Memory and/or multi-layer Neural Networks, and (ii) a Gaussian Process based approach, which relies on the assumption that the channel states are jointly Gaussian. We compare the achieved performances to several benchmarks in two environments: (i) a stochastic environment, and (ii) microcellular outdoor channels obtained by ray-tracing. In general, the deep learning solution shows superior performance in both environments.
Tasks
Published 2019-02-28
URL http://arxiv.org/abs/1902.10890v1
PDF http://arxiv.org/pdf/1902.10890v1.pdf
PWC https://paperswithcode.com/paper/deep-learning-and-gaussian-process-based-band
Repo
Framework

One-Shot Template Matching for Automatic Document Data Capture

Title One-Shot Template Matching for Automatic Document Data Capture
Authors Pranjal Dhakal, Manish Munikar, Bikram Dahal
Abstract In this paper, we propose a novel one-shot template-matching algorithm to automatically capture data from business documents with an aim to minimize manual data entry. Given one annotated document, our algorithm can automatically extract similar data from other documents having the same format. Based on a set of engineered visual and textual features, our method is invariant to changes in position and value. Experiments on a dataset of 595 real invoices demonstrate 86.4% accuracy.
Tasks
Published 2019-10-22
URL https://arxiv.org/abs/1910.10037v1
PDF https://arxiv.org/pdf/1910.10037v1.pdf
PWC https://paperswithcode.com/paper/one-shot-template-matching-for-automatic
Repo
Framework

A Survey of Techniques All Classifiers Can Learn from Deep Networks: Models, Optimizations, and Regularization

Title A Survey of Techniques All Classifiers Can Learn from Deep Networks: Models, Optimizations, and Regularization
Authors Alireza Ghods, Diane J Cook
Abstract Deep neural networks have introduced novel and useful tools to the machine learning community. Other types of classifiers can potentially make use of these tools as well to improve their performance and generality. This paper reviews the current state of the art for deep learning classifier technologies that are being used outside of deep neural networks. Non-network classifiers can employ many components found in deep neural network architectures. In this paper, we review the feature learning, optimization, and regularization methods that form a core of deep network technologies. We then survey non-neural network learning algorithms that make innovative use of these methods to improve classification. Because many opportunities and challenges still exist, we discuss directions that can be pursued to expand the area of deep learning for a variety of classification algorithms.
Tasks
Published 2019-09-10
URL https://arxiv.org/abs/1909.04791v2
PDF https://arxiv.org/pdf/1909.04791v2.pdf
PWC https://paperswithcode.com/paper/techniques-all-classifiers-can-learn-from
Repo
Framework

GRAVITAS: A Model Checking Based Planning and Goal Reasoning Framework for Autonomous Systems

Title GRAVITAS: A Model Checking Based Planning and Goal Reasoning Framework for Autonomous Systems
Authors Hadrien Bride, Jin Song Dong, Ryan Green, Zhe Hou, Brendan Mahony, Martin Oxenham
Abstract While AI techniques have found many successful applications in autonomous systems, many of them permit behaviours that are difficult to interpret and may lead to uncertain results. We follow the “verification as planning” paradigm and propose to use model checking techniques to solve planning and goal reasoning problems for autonomous systems. We give a new formulation of Goal Task Network (GTN) that is tailored for our model checking based framework. We then provide a systematic method that models GTNs in the model checker Process Analysis Toolkit (PAT). We present our planning and goal reasoning system as a framework called Goal Reasoning And Verification for Independent Trusted Autonomous Systems (GRAVITAS) and discuss how it helps provide trustworthy plans in an uncertain environment. Finally, we demonstrate the proposed ideas in an experiment that simulates a survey mission performed by the REMUS-100 autonomous underwater vehicle.
Tasks
Published 2019-10-03
URL https://arxiv.org/abs/1910.01380v1
PDF https://arxiv.org/pdf/1910.01380v1.pdf
PWC https://paperswithcode.com/paper/gravitas-a-model-checking-based-planning-and
Repo
Framework

Quantum Logic Gate Synthesis as a Markov Decision Process

Title Quantum Logic Gate Synthesis as a Markov Decision Process
Authors M. Sohaib Alam
Abstract Reinforcement learning has witnessed recent applications to a variety of tasks in quantum programming. The underlying assumption is that those tasks could be modeled as Markov Decision Processes (MDPs). Here, we investigate the feasibility of this assumption by exploring its consequences for two of the simplest tasks in quantum programming: state preparation and gate compilation. By forming discrete MDPs, focusing exclusively on the single-qubit case, we solve for the optimal policy exactly through policy iteration. We find optimal paths that correspond to the shortest possible sequence of gates to prepare a state, or compile a gate, up to some target accuracy. As an example, we find sequences of H and T gates with length as small as 11 producing ~99% fidelity for states of the form (HT)^{n} 0> with values as large as n=10^{10}. This work provides strong evidence that reinforcement learning can be used for optimal state preparation and gate compilation for larger qubit spaces.
Tasks
Published 2019-12-27
URL https://arxiv.org/abs/1912.12002v1
PDF https://arxiv.org/pdf/1912.12002v1.pdf
PWC https://paperswithcode.com/paper/quantum-logic-gate-synthesis-as-a-markov
Repo
Framework

There is Limited Correlation between Coverage and Robustness for Deep Neural Networks

Title There is Limited Correlation between Coverage and Robustness for Deep Neural Networks
Authors Yizhen Dong, Peixin Zhang, Jingyi Wang, Shuang Liu, Jun Sun, Jianye Hao, Xinyu Wang, Li Wang, Jin Song Dong, Dai Ting
Abstract Deep neural networks (DNN) are increasingly applied in safety-critical systems, e.g., for face recognition, autonomous car control and malware detection. It is also shown that DNNs are subject to attacks such as adversarial perturbation and thus must be properly tested. Many coverage criteria for DNN since have been proposed, inspired by the success of code coverage criteria for software programs. The expectation is that if a DNN is a well tested (and retrained) according to such coverage criteria, it is more likely to be robust. In this work, we conduct an empirical study to evaluate the relationship between coverage, robustness and attack/defense metrics for DNN. Our study is the largest to date and systematically done based on 100 DNN models and 25 metrics. One of our findings is that there is limited correlation between coverage and robustness, i.e., improving coverage does not help improve the robustness. Our dataset and implementation have been made available to serve as a benchmark for future studies on testing DNN.
Tasks Face Recognition, Malware Detection
Published 2019-11-14
URL https://arxiv.org/abs/1911.05904v1
PDF https://arxiv.org/pdf/1911.05904v1.pdf
PWC https://paperswithcode.com/paper/there-is-limited-correlation-between-coverage
Repo
Framework

Selection Heuristics on Semantic Genetic Programming for Classification Problems

Title Selection Heuristics on Semantic Genetic Programming for Classification Problems
Authors Claudia N. Sánchez, Mario Graff
Abstract In a steady-state evolution, tournament selection traditionally uses the fitness function to select the parents, and negative selection chooses an individual to be replaced with an offspring. This contribution focuses on analyzing the behavior, in terms of performance, of different heuristics when used instead of the fitness function in tournament selection. The heuristics analyzed are related to measuring the similarity of the individuals in the semantic space. In addition, the analysis includes random selection and traditional tournament selection. These selection functions were implemented on our Semantic Genetic Programming system, namely EvoDAG, which is inspired by the geometric genetic operators and tested on 30 classification problems with a variable number of samples, variables, and classes. The result indicated that the combination of accuracy and the random selection, in the negative tournament, produces the best combination, and the difference in performances between this combination and the tournament selection is statistically significant. Furthermore, we compare EvoDAG’s performance using the selection heuristics against 18 classifiers that included traditional approaches as well as auto-machine-learning techniques. The results indicate that our proposal is competitive with state-of-art classifiers. Finally, it is worth to mention that EvoDAG is available as open source software.
Tasks
Published 2019-07-16
URL https://arxiv.org/abs/1907.07066v2
PDF https://arxiv.org/pdf/1907.07066v2.pdf
PWC https://paperswithcode.com/paper/selection-heuristics-on-semantic-genetic
Repo
Framework

The Naked Sun: Malicious Cooperation Between Benign-Looking Processes

Title The Naked Sun: Malicious Cooperation Between Benign-Looking Processes
Authors Fabio De Gaspari, Dorjan Hitaj, Giulio Pagnotta, Lorenzo De Carli, Luigi V. Mancini
Abstract Recent progress in machine learning has generated promising results in behavioral malware detection. Behavioral modeling identifies malicious processes via features derived by their runtime behavior. Behavioral features hold great promise as they are intrinsically related to the functioning of each malware, and are therefore considered difficult to evade. Indeed, while a significant amount of results exists on evasion of static malware features, evasion of dynamic features has seen limited work. This paper thoroughly examines the robustness of behavioral malware detectors to evasion, focusing particularly on anti-ransomware evasion. We choose ransomware as its behavior tends to differ significantly from that of benign processes, making it a low-hanging fruit for behavioral detection (and a difficult candidate for evasion). Our analysis identifies a set of novel attacks that distribute the overall malware workload across a small set of cooperating processes to avoid the generation of significant behavioral features. Our most effective attack decreases the accuracy of a state-of-the-art classifier from 98.6% to 0% using only 18 cooperating processes. Furthermore, we show our attacks to be effective against commercial ransomware detectors even in a black-box setting.
Tasks Behavioral Malware Detection, Malware Detection
Published 2019-11-06
URL https://arxiv.org/abs/1911.02423v1
PDF https://arxiv.org/pdf/1911.02423v1.pdf
PWC https://paperswithcode.com/paper/the-naked-sun-malicious-cooperation-between
Repo
Framework

Fast and Deep Graph Neural Networks

Title Fast and Deep Graph Neural Networks
Authors Claudio Gallicchio, Alessio Micheli
Abstract We address the efficiency issue for the construction of a deep graph neural network (GNN). The approach exploits the idea of representing each input graph as a fixed point of a dynamical system (implemented through a recurrent neural network), and leverages a deep architectural organization of the recurrent units. Efficiency is gained by many aspects, including the use of small and very sparse networks, where the weights of the recurrent units are left untrained under the stability condition introduced in this work. This can be viewed as a way to study the intrinsic power of the architecture of a deep GNN, and also to provide insights for the set-up of more complex fully-trained models. Through experimental results, we show that even without training of the recurrent connections, the architecture of small deep GNN is surprisingly able to achieve or improve the state-of-the-art performance on a significant set of tasks in the field of graphs classification.
Tasks
Published 2019-11-20
URL https://arxiv.org/abs/1911.08941v1
PDF https://arxiv.org/pdf/1911.08941v1.pdf
PWC https://paperswithcode.com/paper/fast-and-deep-graph-neural-networks
Repo
Framework

Using learned optimizers to make models robust to input noise

Title Using learned optimizers to make models robust to input noise
Authors Luke Metz, Niru Maheswaranathan, Jonathon Shlens, Jascha Sohl-Dickstein, Ekin D. Cubuk
Abstract State-of-the art vision models can achieve superhuman performance on image classification tasks when testing and training data come from the same distribution. However, when models are tested on corrupted images (e.g. due to scale changes, translations, or shifts in brightness or contrast), performance degrades significantly. Here, we explore the possibility of meta-training a learned optimizer that can train image classification models such that they are robust to common image corruptions. Specifically, we are interested training models that are more robust to noise distributions not present in the training data. We find that a learned optimizer meta-trained to produce models which are robust to Gaussian noise trains models that are more robust to Gaussian noise at other scales compared to traditional optimizers like Adam. The effect of meta-training is more complicated when targeting a more general set of noise distributions, but led to improved performance on half of held-out corruption tasks. Our results suggest that meta-learning provides a novel approach for studying and improving the robustness of deep learning models.
Tasks Image Classification, Meta-Learning
Published 2019-06-08
URL https://arxiv.org/abs/1906.03367v1
PDF https://arxiv.org/pdf/1906.03367v1.pdf
PWC https://paperswithcode.com/paper/using-learned-optimizers-to-make-models
Repo
Framework

Computationally efficient versions of conformal predictive distributions

Title Computationally efficient versions of conformal predictive distributions
Authors Vladimir Vovk, Ivan Petej, Ilia Nouretdinov, Valery Manokhin, Alex Gammerman
Abstract Conformal predictive systems are a recent modification of conformal predictors that output, in regression problems, probability distributions for labels of test observations rather than set predictions. The extra information provided by conformal predictive systems may be useful, e.g., in decision making problems. Conformal predictive systems inherit the relative computational inefficiency of conformal predictors. In this paper we discuss two computationally efficient versions of conformal predictive systems, which we call split conformal predictive systems and cross-conformal predictive systems. The main advantage of split conformal predictive systems is their guaranteed validity, whereas for cross-conformal predictive systems validity only holds empirically and in the absence of excessive randomization. The main advantage of cross-conformal predictive systems is their greater predictive efficiency.
Tasks Decision Making
Published 2019-11-03
URL https://arxiv.org/abs/1911.00941v1
PDF https://arxiv.org/pdf/1911.00941v1.pdf
PWC https://paperswithcode.com/paper/computationally-efficient-versions-of
Repo
Framework
comments powered by Disqus