Paper Group AWR 274
DeepView: Visualizing the behavior of deep neural networks in a part of the data space. Poisoning Attacks with Generative Adversarial Nets. QPyTorch: A Low-Precision Arithmetic Simulation Framework. Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics. Architectural Middleware that Supports Building High-performance, …
DeepView: Visualizing the behavior of deep neural networks in a part of the data space
Title | DeepView: Visualizing the behavior of deep neural networks in a part of the data space |
Authors | Alexander Schulz, Fabian Hinder, Barbara Hammer |
Abstract | Machine learning models using deep architectures have been able to implement increasingly powerful and successful models. However, they also become increasingly more complex, more difficult to comprehend and easier to fool. So far, mostly methods have been proposed to investigate the decision of the model for a single given input datum. In this paper, we propose to visualize a part of the decision function of a deep neural network together with a part of the data set in two dimensions with discriminative dimensionality reduction. This enables us to inspect how different properties of the data are treated by the model, such as multimodality, label noise or biased data. Further, the presented approach is complementary to the mentioned interpretation methods from the literature and hence might be even more useful in combination with those. |
Tasks | Dimensionality Reduction |
Published | 2019-09-19 |
URL | https://arxiv.org/abs/1909.09154v1 |
https://arxiv.org/pdf/1909.09154v1.pdf | |
PWC | https://paperswithcode.com/paper/deepview-visualizing-the-behavior-of-deep |
Repo | https://github.com/LucaHermes/DeepView |
Framework | pytorch |
Poisoning Attacks with Generative Adversarial Nets
Title | Poisoning Attacks with Generative Adversarial Nets |
Authors | Luis Muñoz-González, Bjarne Pfitzner, Matteo Russo, Javier Carnerero-Cano, Emil C. Lupu |
Abstract | Machine learning algorithms are vulnerable to poisoning attacks: An adversary can inject malicious points in the training dataset to influence the learning process and degrade the algorithm’s performance. Optimal poisoning attacks have already been proposed to evaluate worst-case scenarios, modelling attacks as a bi-level optimization problem. Solving these problems is computationally demanding and has limited applicability for some models such as deep networks. In this paper we introduce a novel generative model to craft systematic poisoning attacks against machine learning classifiers generating adversarial training examples, i.e. samples that look like genuine data points but that degrade the classifier’s accuracy when used for training. We propose a Generative Adversarial Net with three components: generator, discriminator, and the target classifier. This approach allows us to model naturally the detectability constrains that can be expected in realistic attacks and to identify the regions of the underlying data distribution that can be more vulnerable to data poisoning. Our experimental evaluation shows the effectiveness of our attack to compromise machine learning classifiers, including deep networks. |
Tasks | data poisoning |
Published | 2019-06-18 |
URL | https://arxiv.org/abs/1906.07773v2 |
https://arxiv.org/pdf/1906.07773v2.pdf | |
PWC | https://paperswithcode.com/paper/poisoning-attacks-with-generative-adversarial |
Repo | https://github.com/lmunoz-gonzalez/Poisoning-Attacks-with-Back-gradient-Optimization |
Framework | none |
QPyTorch: A Low-Precision Arithmetic Simulation Framework
Title | QPyTorch: A Low-Precision Arithmetic Simulation Framework |
Authors | Tianyi Zhang, Zhiqiu Lin, Guandao Yang, Christopher De Sa |
Abstract | Low-precision training reduces computational cost and produces efficient models. Recent research in developing new low-precision training algorithms often relies on simulation to empirically evaluate the statistical effects of quantization while avoiding the substantial overhead of building specific hardware. To support this empirical research, we introduce QPyTorch, a low-precision arithmetic simulation framework. Built natively in PyTorch, QPyTorch provides a convenient interface that minimizes the efforts needed to reliably convert existing codes to study low-precision training. QPyTorch is general, and supports a variety of combinations of precisions, number formats, and rounding options. Additionally, it leverages an efficient fused-kernel approach to reduce simulator overhead, which enables simulation of large-scale, realistic problems. QPyTorch is publicly available at https://github.com/Tiiiger/QPyTorch. |
Tasks | Quantization |
Published | 2019-10-09 |
URL | https://arxiv.org/abs/1910.04540v1 |
https://arxiv.org/pdf/1910.04540v1.pdf | |
PWC | https://paperswithcode.com/paper/qpytorch-a-low-precision-arithmetic |
Repo | https://github.com/Tiiiger/QPyTorch |
Framework | pytorch |
Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics
Title | Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics |
Authors | Yuxin Ma, Tiankai Xie, Jundong Li, Ross Maciejewski |
Abstract | Machine learning models are currently being deployed in a variety of real-world applications where model predictions are used to make decisions about healthcare, bank loans, and numerous other critical tasks. As the deployment of artificial intelligence technologies becomes ubiquitous, it is unsurprising that adversaries have begun developing methods to manipulate machine learning models to their advantage. While the visual analytics community has developed methods for opening the black box of machine learning models, little work has focused on helping the user understand their model vulnerabilities in the context of adversarial attacks. In this paper, we present a visual analytics framework for explaining and exploring model vulnerabilities to adversarial attacks. Our framework employs a multi-faceted visualization scheme designed to support the analysis of data poisoning attacks from the perspective of models, data instances, features, and local structures. We demonstrate our framework through two case studies on binary classifiers and illustrate model vulnerabilities with respect to varying attack strategies. |
Tasks | data poisoning |
Published | 2019-07-17 |
URL | https://arxiv.org/abs/1907.07296v4 |
https://arxiv.org/pdf/1907.07296v4.pdf | |
PWC | https://paperswithcode.com/paper/explaining-vulnerabilities-to-adversarial |
Repo | https://github.com/VADERASU/visual-analytics-adversarial-attacks |
Framework | none |
Architectural Middleware that Supports Building High-performance, Scalable, Ubiquitous, Intelligent Personal Assistants
Title | Architectural Middleware that Supports Building High-performance, Scalable, Ubiquitous, Intelligent Personal Assistants |
Authors | Oscar J. Romero |
Abstract | Intelligent Personal Assistants (IPAs) are software agents that can perform tasks on behalf of individuals and assist them on many of their daily activities. IPAs capabilities are expanding rapidly due to the recent advances on areas such as natural language processing, machine learning, artificial cognition, and ubiquitous computing, which equip the agents with competences to understand what users say, collect information from everyday ubiquitous devices (e.g., smartphones, wearables, tablets, laptops, cars, household appliances, etc.), learn user preferences, deliver data-driven search results, and make decisions based on user’s context. Apart from the inherent complexity of building such IPAs, developers and researchers have to address many critical architectural challenges (e.g., low-latency, scalability, concurrency, ubiquity, code mobility, interoperability, support to cognitive services and reasoning, to name a few.), thereby diverting them from their main goal: building IPAs. Thus, our contribution in this paper is twofold: 1) we propose an architecture for a platform-agnostic, high-performance, ubiquitous, and distributed middleware that alleviates the burdensome task of dealing with low-level implementation details when building IPAs by adding multiple abstraction layers that hide the underlying complexity; and 2) we present an implementation of the middleware that concretizes the aforementioned architecture and allows the development of high-level capabilities while scaling the system up to hundreds of thousands of IPAs with no extra effort. We demonstrate the powerfulness of our middleware by analyzing software metrics for complexity, effort, performance, cohesion and coupling when developing a conversational IPA. |
Tasks | |
Published | 2019-06-05 |
URL | https://arxiv.org/abs/1906.02068v1 |
https://arxiv.org/pdf/1906.02068v1.pdf | |
PWC | https://paperswithcode.com/paper/architectural-middleware-that-supports |
Repo | https://github.com/ojrlopez27/multiuser-framework |
Framework | none |
Pattern Generation Strategies for Improving Recognition of Handwritten Mathematical Expressions
Title | Pattern Generation Strategies for Improving Recognition of Handwritten Mathematical Expressions |
Authors | Anh Duc Le, Bipin Indurkhya, Masaki Nakagawa |
Abstract | Recognition of Handwritten Mathematical Expressions (HMEs) is a challenging problem because of the ambiguity and complexity of two-dimensional handwriting. Moreover, the lack of large training data is a serious issue, especially for academic recognition systems. In this paper, we propose pattern generation strategies that generate shape and structural variations to improve the performance of recognition systems based on a small training set. For data generation, we employ the public databases: CROHME 2014 and 2016 of online HMEs. The first strategy employs local and global distortions to generate shape variations. The second strategy decomposes an online HME into sub-online HMEs to get more structural variations. The hybrid strategy combines both these strategies to maximize shape and structural variations. The generated online HMEs are converted to images for offline HME recognition. We tested our strategies in an end-to-end recognition system constructed from a recent deep learning model: Convolutional Neural Network and attention-based encoder-decoder. The results of experiments on the CROHME 2014 and 2016 databases demonstrate the superiority and effectiveness of our strategies: our hybrid strategy achieved classification rates of 48.78% and 45.60%, respectively, on these databases. These results are competitive compared to others reported in recent literature. Our generated datasets are openly available for research community and constitute a useful resource for the HME recognition research in future. |
Tasks | |
Published | 2019-01-21 |
URL | http://arxiv.org/abs/1901.06763v1 |
http://arxiv.org/pdf/1901.06763v1.pdf | |
PWC | https://paperswithcode.com/paper/pattern-generation-strategies-for-improving |
Repo | https://github.com/ducanh841988/Artificial-Online-Handwritten-Mathematical-Expressions |
Framework | none |
Deep autoregressive models for the efficient variational simulation of many-body quantum systems
Title | Deep autoregressive models for the efficient variational simulation of many-body quantum systems |
Authors | Or Sharir, Yoav Levine, Noam Wies, Giuseppe Carleo, Amnon Shashua |
Abstract | Artificial Neural Networks were recently shown to be an efficient representation of highly-entangled many-body quantum states. In practical applications, neural-network states inherit numerical schemes used in Variational Monte Carlo, most notably the use of Markov-Chain Monte-Carlo (MCMC) sampling to estimate quantum expectations. The local stochastic sampling in MCMC caps the potential advantages of neural networks in two ways: (i) Its intrinsic computational cost sets stringent practical limits on the width and depth of the networks, and therefore limits their expressive capacity; (ii) Its difficulty in generating precise and uncorrelated samples can result in estimations of observables that are very far from their true value. Inspired by the state-of-the-art generative models used in machine learning, we propose a specialized Neural Network architecture that supports efficient and exact sampling, completely circumventing the need for Markov Chain sampling. We demonstrate our approach for two-dimensional interacting spin models, showcasing the ability to obtain accurate results on larger system sizes than those currently accessible to neural-network quantum states. |
Tasks | |
Published | 2019-02-11 |
URL | https://arxiv.org/abs/1902.04057v3 |
https://arxiv.org/pdf/1902.04057v3.pdf | |
PWC | https://paperswithcode.com/paper/deep-autoregressive-models-for-the-efficient |
Repo | https://github.com/HUJI-Deep/FlowKet |
Framework | tf |
Inception-inspired LSTM for Next-frame Video Prediction
Title | Inception-inspired LSTM for Next-frame Video Prediction |
Authors | Matin Hosseini, Anthony S. Maida, Majid Hosseini, Gottumukkala Raju |
Abstract | The problem of video frame prediction has received much interest due to its relevance to many computer vision applications such as autonomous vehicles or robotics. Supervised methods for video frame prediction rely on labeled data, which may not always be available. In this paper, we provide a novel unsupervised deep-learning method called Inception-based LSTM for video frame prediction. The general idea of inception networks is to implement wider networks instead of deeper networks. This network design was shown to improve the performance of image classification. The proposed method is evaluated on both Inception-v1 and Inception-v2 structures. The proposed Inception LSTM methods are compared with convolutional LSTM when applied using PredNet predictive coding framework for both the KITTI and KTH data sets. We observed that the Inception based LSTM outperforms the convolutional LSTM. Also, Inception LSTM has better prediction performance compared to Inception v2 LSTM. However, Inception v2 LSTM has a lower computational cost compared to Inception LSTM. |
Tasks | Autonomous Vehicles, Image Classification, Video Prediction |
Published | 2019-08-28 |
URL | https://arxiv.org/abs/1909.05622v1 |
https://arxiv.org/pdf/1909.05622v1.pdf | |
PWC | https://paperswithcode.com/paper/inception-inspired-lstm-for-next-frame-video |
Repo | https://github.com/matinhosseiny/Inception-inspired-LSTM-for-Video-frame-Prediction |
Framework | none |
Predicting tongue motion in unlabeled ultrasound videos using convolutional LSTM neural network
Title | Predicting tongue motion in unlabeled ultrasound videos using convolutional LSTM neural network |
Authors | Chaojie Zhao, Peng Zhang, Jian Zhu, Chengrui Wu, Huaimin Wang, Kele Xu |
Abstract | A challenge in speech production research is to predict future tongue movements based on a short period of past tongue movements. This study tackles speaker-dependent tongue motion prediction problem in unlabeled ultrasound videos with convolutional long short-term memory (ConvLSTM) networks. The model has been tested on two different ultrasound corpora. ConvLSTM outperforms 3-dimensional convolutional neural network (3DCNN) in predicting the 9\textsuperscript{th} frames based on 8 preceding frames, and also demonstrates good capacity to predict only the tongue contours in future frames. Further tests reveal that ConvLSTM can also learn to predict tongue movements in more distant frames beyond the immediately following frames. Our codes are available at: https://github.com/shuiliwanwu/ConvLstm-ultrasound-videos. |
Tasks | motion prediction |
Published | 2019-02-19 |
URL | http://arxiv.org/abs/1902.06927v1 |
http://arxiv.org/pdf/1902.06927v1.pdf | |
PWC | https://paperswithcode.com/paper/predicting-tongue-motion-in-unlabeled |
Repo | https://github.com/shuiliwanwu/ConvLstm-ultrasound-videos |
Framework | pytorch |
Task-Oriented Optimal Sequencing of Visualization Charts
Title | Task-Oriented Optimal Sequencing of Visualization Charts |
Authors | Danqing Shi, Yang Shi, Xinyue Xu, Nan Chen, Siwei Fu, Hongjin Wu, Nan Cao |
Abstract | A chart sequence is used to describe a series of visualization charts generated in the exploratory analysis by data analysts. It provides information details in each chart as well as a logical relationship among charts. While existing research targets on generating chart sequences that match human’s perceptions, little attention has been paid to formulate task-oriented connections between charts in a chart design space. We present a novel chart sequencing method based on reinforcement learning to capture the connections between charts in the context of three major analysis tasks, including correlation analysis, anomaly detection, and cluster analysis. The proposed method formulates a chart sequencing procedure as an optimization problem, which seeks an optimal policy to sequencing charts for the specific analysis task. In our method, a novel reward function is introduced, which takes both the analysis task and the factor of human cognition into consideration. We conducted one case study and two user studies to evaluate the effectiveness of our method under the application scenarios of visualization demonstration, sequencing charts for reasoning analysis results, and making a chart design choice. The study results showed the power of our method. |
Tasks | Anomaly Detection |
Published | 2019-08-07 |
URL | https://arxiv.org/abs/1908.02502v1 |
https://arxiv.org/pdf/1908.02502v1.pdf | |
PWC | https://paperswithcode.com/paper/task-oriented-optimal-sequencing-of |
Repo | https://github.com/mstaniak/autoEDA-resources |
Framework | none |
An Unsupervised Framework for Comparing Graph Embeddings
Title | An Unsupervised Framework for Comparing Graph Embeddings |
Authors | Bogumil Kaminski, Pawel Pralat, Francois Theberge |
Abstract | Graph embedding is a transformation of vertices of a graph into set of vectors. Good embeddings should capture the graph topology, vertex-to-vertex relationship, and other relevant information about graphs, subgraphs, and vertices. If these objectives are achieved, they are meaningful, understandable, and compressed representations of networks. They also provide more options and tools for data scientists as machine learning on graphs is still quite limited. Finally, vector operations are simpler and faster than comparable operations on graphs. The main challenge is that one needs to make sure that embeddings well describe the properties of the graphs. In particular, the decision has to be made on the embedding dimensionality which highly impacts the quality of an embedding. As a result, selecting the best embedding is a challenging task and very often requires domain experts. In this paper, we propose a ``divergence score’’ that can be assign to various embeddings to distinguish good ones from bad ones. This general framework provides a tool for an unsupervised graph embedding comparison. In order to achieve it, we needed to generalize the well-known Chung-Lu model to incorporate geometry which is interesting on its own rights. In order to test our framework, we did a number of experiments with synthetic networks as well as real-world networks, and various embedding algorithms. | |
Tasks | Graph Embedding |
Published | 2019-05-29 |
URL | https://arxiv.org/abs/1906.04562v1 |
https://arxiv.org/pdf/1906.04562v1.pdf | |
PWC | https://paperswithcode.com/paper/an-unsupervised-framework-for-comparing-graph |
Repo | https://github.com/ftheberge/Comparing_Graph_Embeddings |
Framework | none |
Compositional Fairness Constraints for Graph Embeddings
Title | Compositional Fairness Constraints for Graph Embeddings |
Authors | Avishek Joey Bose, William L. Hamilton |
Abstract | Learning high-quality node embeddings is a key building block for machine learning models that operate on graph data, such as social networks and recommender systems. However, existing graph embedding techniques are unable to cope with fairness constraints, e.g., ensuring that the learned representations do not correlate with certain attributes, such as age or gender. Here, we introduce an adversarial framework to enforce fairness constraints on graph embeddings. Our approach is compositional—meaning that it can flexibly accommodate different combinations of fairness constraints during inference. For instance, in the context of social recommendations, our framework would allow one user to request that their recommendations are invariant to both their age and gender, while also allowing another user to request invariance to just their age. Experiments on standard knowledge graph and recommender system benchmarks highlight the utility of our proposed framework. |
Tasks | Graph Embedding, Recommendation Systems |
Published | 2019-05-25 |
URL | https://arxiv.org/abs/1905.10674v4 |
https://arxiv.org/pdf/1905.10674v4.pdf | |
PWC | https://paperswithcode.com/paper/compositional-fairness-constraints-for-graph |
Repo | https://github.com/joeybose/Flexible-Fairness-Constraints |
Framework | pytorch |
Auto-completion for Data Cells in Relational Tables
Title | Auto-completion for Data Cells in Relational Tables |
Authors | Shuo Zhang, Krisztian Balog |
Abstract | We address the task of auto-completing data cells in relational tables. Such tables describe entities (in rows) with their attributes (in columns). We present the CellAutoComplete framework to tackle several novel aspects of this problem, including: (i) enabling a cell to have multiple, possibly conflicting values, (ii) supplementing the predicted values with supporting evidence, (iii) combining evidence from multiple sources, and (iv) handling the case where a cell should be left empty. Our framework makes use of a large table corpus and a knowledge base as data sources, and consists of preprocessing, candidate value finding, and value ranking components. Using a purpose-built test collection, we show that our approach is 40% more effective than the best baseline. |
Tasks | |
Published | 2019-09-08 |
URL | https://arxiv.org/abs/1909.03443v2 |
https://arxiv.org/pdf/1909.03443v2.pdf | |
PWC | https://paperswithcode.com/paper/auto-completion-for-data-cells-in-relational |
Repo | https://github.com/iai-group/cikm2019-table |
Framework | none |
Speed-up and multi-view extensions to Subclass Discriminant Analysis
Title | Speed-up and multi-view extensions to Subclass Discriminant Analysis |
Authors | Kateryna Chumachenko, Jenni Raitoharju, Alexandros Iosifidis, Moncef Gabbouj |
Abstract | In this paper, we propose a speed-up approach for subclass discriminant analysis and formulate a novel efficient multi-view solution to it. The speed-up approach is developed based on graph embedding and spectral regression approaches that involve eigendecomposition of the corresponding Laplacian matrix and regression to its eigenvectors. We show that by exploiting the structure of the between-class Laplacian matrix, the eigendecomposition step can be substituted with a much faster process. Furthermore, we formulate a novel criterion for multi-view subclass discriminant analysis and show that an efficient solution for it can be obtained in a similar to the single-view manner. We evaluate the proposed methods on nine single-view and nine multi-view datasets and compare them with related existing approaches. Experimental results show that the proposed solutions achieve competitive performance, often outperforming the existing methods. At the same time, they significantly decrease the training time. |
Tasks | Graph Embedding |
Published | 2019-05-02 |
URL | https://arxiv.org/abs/1905.00794v1 |
https://arxiv.org/pdf/1905.00794v1.pdf | |
PWC | https://paperswithcode.com/paper/speed-up-and-multi-view-extensions-to |
Repo | https://github.com/katerynaCh/fastSDA |
Framework | none |
Probabilistic Models with Deep Neural Networks
Title | Probabilistic Models with Deep Neural Networks |
Authors | Andrés R. Masegosa, Rafael Cabañas, Helge Langseth, Thomas D. Nielsen, Antonio Salmerón |
Abstract | Recent advances in statistical inference have significantly expanded the toolbox of probabilistic modeling. Historically, probabilistic modeling has been constrained to (i) very restricted model classes where exact or approximate probabilistic inference were feasible, and (ii) small or medium-sized data sets which fit within the main memory of the computer. However, developments in variational inference, a general form of approximate probabilistic inference originated in statistical physics, are allowing probabilistic modeling to overcome these restrictions: (i) Approximate probabilistic inference is now possible over a broad class of probabilistic models containing a large number of parameters, and (ii) scalable inference methods based on stochastic gradient descent and distributed computation engines allow to apply probabilistic modeling over massive data sets. One important practical consequence of these advances is the possibility to include deep neural networks within a probabilistic model to capture complex non-linear stochastic relationships between random variables. These advances in conjunction with the release of novel probabilistic modeling toolboxes have greatly expanded the scope of application of probabilistic models, and allow these models to take advantage of the recent strides made by the deep learning community. In this paper we review the main concepts, methods and tools needed to use deep neural networks within a probabilistic modeling framework. |
Tasks | |
Published | 2019-08-09 |
URL | https://arxiv.org/abs/1908.03442v3 |
https://arxiv.org/pdf/1908.03442v3.pdf | |
PWC | https://paperswithcode.com/paper/probabilistic-models-with-deep-neural |
Repo | https://github.com/PGM-Lab/ProbModelsDNNs |
Framework | tf |