January 31, 2020

2971 words 14 mins read

Paper Group ANR 163

Paper Group ANR 163

Proceedings of the 4th Workshop on Formal Reasoning about Causation, Responsibility, and Explanations in Science and Technology. Fast Neural Network Verification via Shadow Prices. Face Recognition System. Component Mismatches Are a Critical Bottleneck to Fielding AI-Enabled Systems in the Public Sector. Adversarial Examples for Electrocardiograms. …

Proceedings of the 4th Workshop on Formal Reasoning about Causation, Responsibility, and Explanations in Science and Technology

Title Proceedings of the 4th Workshop on Formal Reasoning about Causation, Responsibility, and Explanations in Science and Technology
Authors Georgiana Caltais, Jean Krivine
Abstract The fourth edition of the international workshop on Causation, Responsibility and Explanation took place in Prague (Czech Republic) as part of ETAPS 2019. The program consisted in 5 invited speakers and 4 regular papers, whose selection was based on a careful reviewing process and that are included in these proceedings.
Tasks
Published 2019-10-30
URL https://arxiv.org/abs/1910.13641v1
PDF https://arxiv.org/pdf/1910.13641v1.pdf
PWC https://paperswithcode.com/paper/proceedings-of-the-4th-workshop-on-formal
Repo
Framework

Fast Neural Network Verification via Shadow Prices

Title Fast Neural Network Verification via Shadow Prices
Authors Vicenc Rubies Royo, Roberto Calandra, Dusan M. Stipanovic, Claire Tomlin
Abstract To use neural networks in safety-critical settings it is paramount to provide assurances on their runtime operation. Recent work on ReLU networks has sought to verify whether inputs belonging to a bounded box can ever yield some undesirable output. Input-splitting procedures, a particular type of verification mechanism, do so by recursively partitioning the input set into smaller sets. The efficiency of these methods is largely determined by the number of splits the box must undergo before the property can be verified. In this work, we propose a new technique based on shadow prices that fully exploits the information of the problem yielding a more efficient generation of splits than the state-of-the-art. Results on the Airborne Collision Avoidance System (ACAS) benchmark verification tasks show a considerable reduction in the partitions generated which substantially reduces computation times. These results open the door to improved verification methods for a wide variety of machine learning applications including vision and control.
Tasks
Published 2019-02-19
URL http://arxiv.org/abs/1902.07247v2
PDF http://arxiv.org/pdf/1902.07247v2.pdf
PWC https://paperswithcode.com/paper/fast-neural-network-verification-via-shadow
Repo
Framework

Face Recognition System

Title Face Recognition System
Authors Yang Li, Sangwhan Cha
Abstract Deep learning is one of the new and important branches in machine learning. Deep learning refers to a set of algorithms that solve various problems such as images and texts by using various machine learning algorithms in multi-layer neural networks. Deep learning can be classified as a neural network from the general category, but there are many changes in the concrete realization. At the core of deep learning is feature learning, which is designed to obtain hierarchical information through hierarchical networks, so as to solve the important problems that previously required artificial design features. Deep Learning is a framework that contains several important algorithms. For different applications (images, voice, text), you need to use different network models to achieve better results. With the development of deep learning and the introduction of deep convolutional neural networks, the accuracy and speed of face recognition have made great strides. However, as we said above, the results from different networks and models are very different. In this paper, facial features are extracted by merging and comparing multiple models, and then a deep neural network is constructed to train and construct the combined features. In this way, the advantages of multiple models can be combined to mention the recognition accuracy. After getting a model with high accuracy, we build a product model. This article compares the pure-client model with the server-client model, analyzes the pros and cons of the two models, and analyzes the various commercial products that are required for the server-client model.
Tasks Face Recognition
Published 2019-01-08
URL http://arxiv.org/abs/1901.02452v1
PDF http://arxiv.org/pdf/1901.02452v1.pdf
PWC https://paperswithcode.com/paper/face-recognition-system
Repo
Framework

Component Mismatches Are a Critical Bottleneck to Fielding AI-Enabled Systems in the Public Sector

Title Component Mismatches Are a Critical Bottleneck to Fielding AI-Enabled Systems in the Public Sector
Authors Grace A. Lewis, Stephany Bellomo, April Galyardt
Abstract The use of machine learning or artificial intelligence (ML/AI) holds substantial potential toward improving many functions and needs of the public sector. In practice however, integrating ML/AI components into public sector applications is severely limited not only by the fragility of these components and their algorithms, but also because of mismatches between components of ML-enabled systems. For example, if an ML model is trained on data that is different from data in the operational environment, field performance of the ML component will be dramatically reduced. Separate from software engineering considerations, the expertise needed to field an ML/AI component within a system frequently comes from outside software engineering. As a result, assumptions and even descriptive language used by practitioners from these different disciplines can exacerbate other challenges to integrating ML/AI components into larger systems. We are investigating classes of mismatches in ML/AI systems integration, to identify the implicit assumptions made by practitioners in different fields (data scientists, software engineers, operations staff) and find ways to communicate the appropriate information explicitly. We will discuss a few categories of mismatch, and provide examples from each class. To enable ML/AI components to be fielded in a meaningful way, we will need to understand the mismatches that exist and develop practices to mitigate the impacts of these mismatches.
Tasks
Published 2019-10-14
URL https://arxiv.org/abs/1910.06136v1
PDF https://arxiv.org/pdf/1910.06136v1.pdf
PWC https://paperswithcode.com/paper/component-mismatches-are-a-critical
Repo
Framework

Adversarial Examples for Electrocardiograms

Title Adversarial Examples for Electrocardiograms
Authors Xintian Han, Yuxuan Hu, Luca Foschini, Larry Chinitz, Lior Jankelson, Rajesh Ranganath
Abstract In recent years, the electrocardiogram (ECG) has seen a large diffusion in both medical and commercial applications, fueled by the rise of single-lead versions. Single-lead ECG can be embedded in medical devices and wearable products such as the injectable Medtronic Linq monitor, the iRhythm Ziopatch wearable monitor, and the Apple Watch Series 4. Recently, deep neural networks have been used to automatically analyze ECG tracings, outperforming even physicians specialized in cardiac electrophysiology in detecting certain rhythm irregularities. However, deep learning classifiers have been shown to be brittle to adversarial examples, which are examples created to look incontrovertibly belonging to a certain class to a human eye but contain subtle features that fool the classifier into misclassifying them into the wrong class. Very recently, adversarial examples have also been created for medical-related tasks. Yet, traditional attack methods to create adversarial examples, such as projected gradient descent (PGD) do not extend directly to ECG signals, as they generate examples that introduce square wave artifacts that are not physiologically plausible. Here, we developed a method to construct smoothed adversarial examples for single-lead ECG. First, we implemented a neural network model achieving state-of-the-art performance on the data from the 2017 PhysioNet/Computing-in-Cardiology Challenge for arrhythmia detection from single lead ECG classification. For this model, we utilized a new technique to generate smoothed examples to produce signals that are 1) indistinguishable to cardiologists from the original examples and 2) incorrectly classified by the neural network. Finally, we show that adversarial examples are not unique and provide a general technique to collate and perturb known adversarial examples to create new ones.
Tasks Adversarial Defense, Arrhythmia Detection, ECG Classification, Electrocardiography (ECG)
Published 2019-05-13
URL https://arxiv.org/abs/1905.05163v2
PDF https://arxiv.org/pdf/1905.05163v2.pdf
PWC https://paperswithcode.com/paper/adversarial-examples-for-electrocardiograms
Repo
Framework

ViGGO: A Video Game Corpus for Data-To-Text Generation in Open-Domain Conversation

Title ViGGO: A Video Game Corpus for Data-To-Text Generation in Open-Domain Conversation
Authors Juraj Juraska, Kevin K. Bowden, Marilyn Walker
Abstract The uptake of deep learning in natural language generation (NLG) led to the release of both small and relatively large parallel corpora for training neural models. The existing data-to-text datasets are, however, aimed at task-oriented dialogue systems, and often thus limited in diversity and versatility. They are typically crowdsourced, with much of the noise left in them. Moreover, current neural NLG models do not take full advantage of large training data, and due to their strong generalizing properties produce sentences that look template-like regardless. We therefore present a new corpus of 7K samples, which (1) is clean despite being crowdsourced, (2) has utterances of 9 generalizable and conversational dialogue act types, making it more suitable for open-domain dialogue systems, and (3) explores the domain of video games, which is new to dialogue systems despite having excellent potential for supporting rich conversations.
Tasks Data-to-Text Generation, Task-Oriented Dialogue Systems, Text Generation
Published 2019-10-26
URL https://arxiv.org/abs/1910.12129v1
PDF https://arxiv.org/pdf/1910.12129v1.pdf
PWC https://paperswithcode.com/paper/viggo-a-video-game-corpus-for-data-to-text
Repo
Framework

Ultra Low-Power and Real-time ECG Classification Based on STDP and R-STDP Neural Networks for Wearable Devices

Title Ultra Low-Power and Real-time ECG Classification Based on STDP and R-STDP Neural Networks for Wearable Devices
Authors Alireza Amirshahi, Matin Hashemi
Abstract This paper presents a novel ECG classification algorithm for real-time cardiac monitoring on ultra low-power wearable devices. The proposed solution is based on spiking neural networks which are the third generation of neural networks. In specific, we employ spike-timing dependent plasticity (STDP), and reward-modulated STDP (R-STDP), in which the model weights are trained according to the timings of spike signals, and reward or punishment signals. Experiments show that the proposed solution is suitable for real-time operation, achieves comparable accuracy with respect to previous methods, and more importantly, its energy consumption is significantly smaller than previous neural network based solutions.
Tasks ECG Classification
Published 2019-05-08
URL https://arxiv.org/abs/1905.02954v4
PDF https://arxiv.org/pdf/1905.02954v4.pdf
PWC https://paperswithcode.com/paper/ultra-low-power-and-real-time-ecg-1
Repo
Framework

Deep Time-Frequency Representation and Progressive Decision Fusion for ECG Classification

Title Deep Time-Frequency Representation and Progressive Decision Fusion for ECG Classification
Authors Jing Zhang, Jing Tian, Yang Cao, Yuxiang Yang, Xiaobin Xu
Abstract Early recognition of abnormal rhythms in ECG signals is crucial for monitoring and diagnosing patients’ cardiac conditions, increasing the success rate of the treatment. Classifying abnormal rhythms into exact categories is very challenging due to the broad taxonomy of rhythms, noises and lack of large-scale real-world annotated data. Different from previous methods that utilize hand-crafted features or learn features from the original signal domain, we propose a novel ECG classification method by learning deep time-frequency representation and progressive decision fusion at different temporal scales in an end-to-end manner. First, the ECG wave signal is transformed into the time-frequency domain by using the Short-Time Fourier Transform. Next, several scale-specific deep convolutional neural networks are trained on ECG samples of a specific length. Finally, a progressive online decision fusion method is proposed to fuse decisions from the scale-specific models into a more accurate and stable one. Extensive experiments on both synthetic and real-world ECG datasets demonstrate the effectiveness and efficiency of the proposed method.
Tasks ECG Classification
Published 2019-01-19
URL https://arxiv.org/abs/1901.06469v3
PDF https://arxiv.org/pdf/1901.06469v3.pdf
PWC https://paperswithcode.com/paper/fine-grained-ecg-classification-based-on-deep
Repo
Framework

Constant Time Graph Neural Networks

Title Constant Time Graph Neural Networks
Authors Ryoma Sato, Makoto Yamada, Hisashi Kashima
Abstract The recent advancements in graph neural networks (GNNs) have led to state-of-the-art performances in various applications, including chemo-informatics, question-answering systems, and recommender systems. However, scaling up these methods to huge graphs, such as social networks and Web graphs, remains a challenge. In particular, the existing methods for accelerating GNNs either are not theoretically guaranteed in terms of the approximation error or incur at least a linear time computation cost. In this study, we reveal the query complexity of the uniform node sampling scheme for Message Passing Neural Networks including GraphSAGE, graph attention networks (GATs), and graph convolutional networks (GCNs). Surprisingly, our analysis reveals that the complexity of the node sampling method is completely independent of the number of the nodes, edges, and neighbors of the input and depends only on the error tolerance and confidence probability while providing a theoretical guarantee for the approximation error. To the best of our knowledge, this is the first paper to provide a theoretical guarantee of approximation for GNNs within constant time. Through experiments with synthetic and real-world datasets, we investigated the speed and precision of the node sampling scheme and validated our theoretical results.
Tasks Question Answering, Recommendation Systems
Published 2019-01-23
URL https://arxiv.org/abs/1901.07868v3
PDF https://arxiv.org/pdf/1901.07868v3.pdf
PWC https://paperswithcode.com/paper/constant-time-graph-neural-networks
Repo
Framework

Distributed representation of multi-sense words: A loss-driven approach

Title Distributed representation of multi-sense words: A loss-driven approach
Authors Saurav Manchanda, George Karypis
Abstract Word2Vec’s Skip Gram model is the current state-of-the-art approach for estimating the distributed representation of words. However, it assumes a single vector per word, which is not well-suited for representing words that have multiple senses. This work presents LDMI, a new model for estimating distributional representations of words. LDMI relies on the idea that, if a word carries multiple senses, then having a different representation for each of its senses should lead to a lower loss associated with predicting its co-occurring words, as opposed to the case when a single vector representation is used for all the senses. After identifying the multi-sense words, LDMI clusters the occurrences of these words to assign a sense to each occurrence. Experiments on the contextual word similarity task show that LDMI leads to better performance than competing approaches.
Tasks
Published 2019-04-14
URL http://arxiv.org/abs/1904.06725v1
PDF http://arxiv.org/pdf/1904.06725v1.pdf
PWC https://paperswithcode.com/paper/distributed-representation-of-multi-sense
Repo
Framework

Targeted Source Detection for Environmental Data

Title Targeted Source Detection for Environmental Data
Authors Guanjie Zheng, Mengqi Liu, Tao Wen, Hongjian Wang, Huaxiu Yao, Susan L. Brantley, Zhenhui Li
Abstract In the face of growing needs for water and energy, a fundamental understanding of the environmental impacts of human activities becomes critical for managing water and energy resources, remedying water pollution, and making regulatory policy wisely. Among activities that impact the environment, oil and gas production, wastewater transport, and urbanization are included. In addition to the occurrence of anthropogenic contamination, the presence of some contaminants (e.g., methane, salt, and sulfate) of natural origin is not uncommon. Therefore, scientists sometimes find it difficult to identify the sources of contaminants in the coupled natural and human systems. In this paper, we propose a technique to simultaneously conduct source detection and prediction, which outperforms other approaches in the interdisciplinary case study of the identification of potential groundwater contamination within a region of high-density shale gas development.
Tasks
Published 2019-08-29
URL https://arxiv.org/abs/1908.11056v1
PDF https://arxiv.org/pdf/1908.11056v1.pdf
PWC https://paperswithcode.com/paper/targeted-source-detection-for-environmental
Repo
Framework

Numeric Solutions of Eigenvalue Problems for Generic Nonlinear Operators

Title Numeric Solutions of Eigenvalue Problems for Generic Nonlinear Operators
Authors Ester Hait-Fraenkel, Guy Gilboa
Abstract Numerical methods for solving linear eigenvalue problem are widely studiedand used in science and engineering. In this paper, we propose a generalizednumerical method for solving eigenproblems for generic, nonlinear opera-tors. This has potentially wide implications, since most image processingalgorithms (e.g. denoising) can be viewed as nonlinear operators, whoseeigenproblem analysis provides information on the most- and least-suitablefunctions as input. We solve the problem by a nonlinear adaptation of thepower method, a well known linear eigensolver. An analysis and valida-tion framework is proposed, as well as preliminary theory. We validate themethod using total-variation (TV) and demonstrate it on the EPLL denoiser(Zoran-Weiss). Finally, we suggest an encryption-decryption application.
Tasks Denoising
Published 2019-09-12
URL https://arxiv.org/abs/1909.12775v2
PDF https://arxiv.org/pdf/1909.12775v2.pdf
PWC https://paperswithcode.com/paper/numeric-solutions-of-eigenvalue-problems-for
Repo
Framework

Let’s measure run time! Extending the IR replicability infrastructure to include performance aspects

Title Let’s measure run time! Extending the IR replicability infrastructure to include performance aspects
Authors Sebastian Hofstätter, Allan Hanbury
Abstract Establishing a docker-based replicability infrastructure offers the community a great opportunity: measuring the run time of information retrieval systems. The time required to present query results to a user is paramount to the users satisfaction. Recent advances in neural IR re-ranking models put the issue of query latency at the forefront. They bring a complex trade-off between performance and effectiveness based on a myriad of factors: the choice of encoding model, network architecture, hardware acceleration and many others. The best performing models (currently using the BERT transformer model) run orders of magnitude more slowly than simpler architectures. We aim to broaden the focus of the neural IR community to include performance considerations – to sustain the practical applicability of our innovations. In this position paper we supply our argument with a case study exploring the performance of different neural re-ranking models. Finally, we propose to extend the OSIRRC docker-based replicability infrastructure with two performance focused benchmark scenarios.
Tasks Information Retrieval
Published 2019-07-10
URL https://arxiv.org/abs/1907.04614v1
PDF https://arxiv.org/pdf/1907.04614v1.pdf
PWC https://paperswithcode.com/paper/lets-measure-run-time-extending-the-ir
Repo
Framework

Minimax Optimal Algorithms for Adversarial Bandit Problem with Multiple Plays

Title Minimax Optimal Algorithms for Adversarial Bandit Problem with Multiple Plays
Authors N. Mert Vural, Hakan Gokcesu, Kaan Gokcesu, Suleyman S. Kozat
Abstract We investigate the adversarial bandit problem with multiple plays under semi-bandit feedback. We introduce a highly efficient algorithm that asymptotically achieves the performance of the best switching $m$-arm strategy with minimax optimal regret bounds. To construct our algorithm, we introduce a new expert advice algorithm for the multiple-play setting. By using our expert advice algorithm, we additionally improve the best-known high-probability bound for the multi-play setting by $O(\sqrt{m})$. Our results are guaranteed to hold in an individual sequence manner since we have no statistical assumption on the bandit arm gains. Through an extensive set of experiments involving synthetic and real data, we demonstrate significant performance gains achieved by the proposed algorithm with respect to the state-of-the-art algorithms.
Tasks
Published 2019-11-25
URL https://arxiv.org/abs/1911.11122v1
PDF https://arxiv.org/pdf/1911.11122v1.pdf
PWC https://paperswithcode.com/paper/minimax-optimal-algorithms-for-adversarial
Repo
Framework

On the Estimation and Use of Statistical Modelling in Information Retrieval

Title On the Estimation and Use of Statistical Modelling in Information Retrieval
Authors Casper Petersen
Abstract Several tasks in information retrieval (IR) rely on assumptions regarding the distribution of some property (such as term frequency) in the data being processed. This thesis argues that such distributional assumptions can lead to incorrect conclusions and proposes a statistically principled method for determining the “true” distribution. This thesis further applies this method to derive a new family of ranking models that adapt their computations to the statistics of the data being processed. Experimental evaluation shows results on par or better than multiple strong baselines on several TREC collections. Overall, this thesis concludes that distributional assumptions can be replaced with an effective, efficient and principled method for determining the “true” distribution and that using the “true” distribution can lead to improved retrieval performance.
Tasks Information Retrieval
Published 2019-03-30
URL http://arxiv.org/abs/1904.00289v1
PDF http://arxiv.org/pdf/1904.00289v1.pdf
PWC https://paperswithcode.com/paper/on-the-estimation-and-use-of-statistical
Repo
Framework
comments powered by Disqus