Paper Group AWR 94
Learning the Joint Representation of Heterogeneous Temporal Events for Clinical Endpoint Prediction. Capsule-Forensics: Using Capsule Networks to Detect Forged Images and Videos. Integral Privacy for Sampling. A Reductions Approach to Fair Classification. MCRM: Mother Compact Recurrent Memory. Seq2Seq-Vis: A Visual Debugging Tool for Sequence-to-Se …
Learning the Joint Representation of Heterogeneous Temporal Events for Clinical Endpoint Prediction
Title | Learning the Joint Representation of Heterogeneous Temporal Events for Clinical Endpoint Prediction |
Authors | Luchen Liu, Jianhao Shen, Ming Zhang, Zichang Wang, Jian Tang |
Abstract | The availability of a large amount of electronic health records (EHR) provides huge opportunities to improve health care service by mining these data. One important application is clinical endpoint prediction, which aims to predict whether a disease, a symptom or an abnormal lab test will happen in the future according to patients’ history records. This paper develops deep learning techniques for clinical endpoint prediction, which are effective in many practical applications. However, the problem is very challenging since patients’ history records contain multiple heterogeneous temporal events such as lab tests, diagnosis, and drug administrations. The visiting patterns of different types of events vary significantly, and there exist complex nonlinear relationships between different events. In this paper, we propose a novel model for learning the joint representation of heterogeneous temporal events. The model adds a new gate to control the visiting rates of different events which effectively models the irregular patterns of different events and their nonlinear correlations. Experiment results with real-world clinical data on the tasks of predicting death and abnormal lab tests prove the effectiveness of our proposed approach over competitive baselines. |
Tasks | |
Published | 2018-03-13 |
URL | http://arxiv.org/abs/1803.04837v4 |
http://arxiv.org/pdf/1803.04837v4.pdf | |
PWC | https://paperswithcode.com/paper/learning-the-joint-representation-of |
Repo | https://github.com/pkusjh/HELSTM |
Framework | none |
Capsule-Forensics: Using Capsule Networks to Detect Forged Images and Videos
Title | Capsule-Forensics: Using Capsule Networks to Detect Forged Images and Videos |
Authors | Huy H. Nguyen, Junichi Yamagishi, Isao Echizen |
Abstract | Recent advances in media generation techniques have made it easier for attackers to create forged images and videos. State-of-the-art methods enable the real-time creation of a forged version of a single video obtained from a social network. Although numerous methods have been developed for detecting forged images and videos, they are generally targeted at certain domains and quickly become obsolete as new kinds of attacks appear. The method introduced in this paper uses a capsule network to detect various kinds of spoofs, from replay attacks using printed images or recorded videos to computer-generated videos using deep convolutional neural networks. It extends the application of capsule networks beyond their original intention to the solving of inverse graphics problems. |
Tasks | Detect Forged Images And Videos |
Published | 2018-10-26 |
URL | http://arxiv.org/abs/1810.11215v1 |
http://arxiv.org/pdf/1810.11215v1.pdf | |
PWC | https://paperswithcode.com/paper/capsule-forensics-using-capsule-networks-to |
Repo | https://github.com/nii-yamagishilab/Capsule-Forensics |
Framework | pytorch |
Integral Privacy for Sampling
Title | Integral Privacy for Sampling |
Authors | Hisham Husain, Zac Cranko, Richard Nock |
Abstract | Differential privacy is a leading protection setting, focused by design on individual privacy. Many applications, in medical / pharmaceutical domains or social networks, rather posit privacy at a group level, a setting we call integral privacy. We aim for the strongest form of privacy: the group size is in particular not known in advance. We study a problem with related applications in domains cited above that have recently met with substantial recent press: sampling. Keeping correct utility levels in such a strong model of statistical indistinguishability looks difficult to be achieved with the usual differential privacy toolbox because it would typically scale in the worst case the sensitivity by the sample size and so the noise variance by up to its square. We introduce a trick specific to sampling that bypasses the sensitivity analysis. Privacy enforces an information theoretic barrier on approximation, and we show how to reach this barrier with guarantees on the approximation of the target non private density. We do so using a recent approach to non private density estimation relying on the original boosting theory, learning the sufficient statistics of an exponential family with classifiers. Approximation guarantees cover the mode capture problem. In the context of learning, the sampling problem is particularly important: because integral privacy enjoys the same closure under post-processing as differential privacy does, any algorithm using integrally privacy sampled data would result in an output equally integrally private. We also show that this brings fairness guarantees on post-processing that would eventually elude classical differential privacy: any decision process has bounded data-dependent bias when the data is integrally privately sampled. Experimental results against private kernel density estimation and private GANs displays the quality of our results. |
Tasks | Density Estimation |
Published | 2018-06-13 |
URL | https://arxiv.org/abs/1806.04819v5 |
https://arxiv.org/pdf/1806.04819v5.pdf | |
PWC | https://paperswithcode.com/paper/integral-privacy-for-sampling-from-mollifier |
Repo | https://github.com/karokaram/PrivatedBoostedDensities |
Framework | none |
A Reductions Approach to Fair Classification
Title | A Reductions Approach to Fair Classification |
Authors | Alekh Agarwal, Alina Beygelzimer, Miroslav Dudík, John Langford, Hanna Wallach |
Abstract | We present a systematic approach for achieving fairness in a binary classification setting. While we focus on two well-known quantitative definitions of fairness, our approach encompasses many other previously studied definitions as special cases. The key idea is to reduce fair classification to a sequence of cost-sensitive classification problems, whose solutions yield a randomized classifier with the lowest (empirical) error subject to the desired constraints. We introduce two reductions that work for any representation of the cost-sensitive classifier and compare favorably to prior baselines on a variety of data sets, while overcoming several of their disadvantages. |
Tasks | |
Published | 2018-03-06 |
URL | http://arxiv.org/abs/1803.02453v3 |
http://arxiv.org/pdf/1803.02453v3.pdf | |
PWC | https://paperswithcode.com/paper/a-reductions-approach-to-fair-classification |
Repo | https://github.com/Microsoft/fairlearn |
Framework | none |
MCRM: Mother Compact Recurrent Memory
Title | MCRM: Mother Compact Recurrent Memory |
Authors | Abduallah A. Mohamed, Christian Claudel |
Abstract | LSTMs and GRUs are the most common recurrent neural network architectures used to solve temporal sequence problems. The two architectures have differing data flows dealing with a common component called the cell state (also referred to as the memory). We attempt to enhance the memory by presenting a modification that we call the Mother Compact Recurrent Memory (MCRM). MCRMs are a type of a nested LSTM-GRU architecture where the cell state is the GRU hidden state. The concatenation of the forget gate and input gate interactions from the LSTM are considered an input to the GRU cell. Because MCRMs has this type of nesting, MCRMs have a compact memory pattern consisting of neurons that acts explicitly in both long-term and short-term fashions. For some specific tasks, empirical results show that MCRMs outperform previously used architectures. |
Tasks | |
Published | 2018-08-04 |
URL | https://arxiv.org/abs/1808.02016v3 |
https://arxiv.org/pdf/1808.02016v3.pdf | |
PWC | https://paperswithcode.com/paper/mcrm-mother-compact-recurrent-memory |
Repo | https://github.com/abduallahmohamed/MCRM |
Framework | pytorch |
Seq2Seq-Vis: A Visual Debugging Tool for Sequence-to-Sequence Models
Title | Seq2Seq-Vis: A Visual Debugging Tool for Sequence-to-Sequence Models |
Authors | Hendrik Strobelt, Sebastian Gehrmann, Michael Behrisch, Adam Perer, Hanspeter Pfister, Alexander M. Rush |
Abstract | Neural Sequence-to-Sequence models have proven to be accurate and robust for many sequence prediction tasks, and have become the standard approach for automatic translation of text. The models work in a five stage blackbox process that involves encoding a source sequence to a vector space and then decoding out to a new target sequence. This process is now standard, but like many deep learning methods remains quite difficult to understand or debug. In this work, we present a visual analysis tool that allows interaction with a trained sequence-to-sequence model through each stage of the translation process. The aim is to identify which patterns have been learned and to detect model errors. We demonstrate the utility of our tool through several real-world large-scale sequence-to-sequence use cases. |
Tasks | |
Published | 2018-04-25 |
URL | http://arxiv.org/abs/1804.09299v2 |
http://arxiv.org/pdf/1804.09299v2.pdf | |
PWC | https://paperswithcode.com/paper/seq2seq-vis-a-visual-debugging-tool-for |
Repo | https://github.com/HendrikStrobelt/Seq2Seq-Vis |
Framework | none |
Graph Convolutional Reinforcement Learning
Title | Graph Convolutional Reinforcement Learning |
Authors | Jiechuan Jiang, Chen Dun, Tiejun Huang, Zongqing Lu |
Abstract | Learning to cooperate is crucially important in multi-agent environments. The key is to understand the mutual interplay between agents. However, multi-agent environments are highly dynamic, where agents keep moving and their neighbors change quickly. This makes it hard to learn abstract representations of mutual interplay between agents. To tackle these difficulties, we propose graph convolutional reinforcement learning, where graph convolution adapts to the dynamics of the underlying graph of the multi-agent environment, and relation kernels capture the interplay between agents by their relation representations. Latent features produced by convolutional layers from gradually increased receptive fields are exploited to learn cooperation, and cooperation is further improved by temporal relation regularization for consistency. Empirically, we show that our method substantially outperforms existing methods in a variety of cooperative scenarios. |
Tasks | Decision Making |
Published | 2018-10-22 |
URL | https://arxiv.org/abs/1810.09202v5 |
https://arxiv.org/pdf/1810.09202v5.pdf | |
PWC | https://paperswithcode.com/paper/graph-convolutional-reinforcement-learning |
Repo | https://github.com/PKU-AI-Edge/DGN |
Framework | tf |
Evolutionary Data Measures: Understanding the Difficulty of Text Classification Tasks
Title | Evolutionary Data Measures: Understanding the Difficulty of Text Classification Tasks |
Authors | Edward Collins, Nikolai Rozanov, Bingbing Zhang |
Abstract | Classification tasks are usually analysed and improved through new model architectures or hyperparameter optimisation but the underlying properties of datasets are discovered on an ad-hoc basis as errors occur. However, understanding the properties of the data is crucial in perfecting models. In this paper we analyse exactly which characteristics of a dataset best determine how difficult that dataset is for the task of text classification. We then propose an intuitive measure of difficulty for text classification datasets which is simple and fast to calculate. We show that this measure generalises to unseen data by comparing it to state-of-the-art datasets and results. This measure can be used to analyse the precise source of errors in a dataset and allows fast estimation of how difficult a dataset is to learn. We searched for this measure by training 12 classical and neural network based models on 78 real-world datasets, then use a genetic algorithm to discover the best measure of difficulty. Our difficulty-calculating code ( https://github.com/Wluper/edm ) and datasets ( http://data.wluper.com ) are publicly available. |
Tasks | Text Classification |
Published | 2018-11-05 |
URL | http://arxiv.org/abs/1811.01910v2 |
http://arxiv.org/pdf/1811.01910v2.pdf | |
PWC | https://paperswithcode.com/paper/evolutionary-data-measures-understanding-the |
Repo | https://github.com/Wluper/edm |
Framework | none |
Deep Learning on Graphs: A Survey
Title | Deep Learning on Graphs: A Survey |
Authors | Ziwei Zhang, Peng Cui, Wenwu Zhu |
Abstract | Deep learning has been shown to be successful in a number of domains, ranging from acoustics, images, to natural language processing. However, applying deep learning to the ubiquitous graph data is non-trivial because of the unique characteristics of graphs. Recently, substantial research efforts have been devoted to applying deep learning methods to graphs, resulting in beneficial advances in graph analysis techniques. In this survey, we comprehensively review the different types of deep learning methods on graphs. We divide the existing methods into five categories based on their model architectures and training strategies: graph recurrent neural networks, graph convolutional networks, graph autoencoders, graph reinforcement learning, and graph adversarial methods. We then provide a comprehensive overview of these methods in a systematic manner mainly by following their development history. We also analyze the differences and compositions of different methods. Finally, we briefly outline the applications in which they have been used and discuss potential future research directions. |
Tasks | |
Published | 2018-12-11 |
URL | https://arxiv.org/abs/1812.04202v3 |
https://arxiv.org/pdf/1812.04202v3.pdf | |
PWC | https://paperswithcode.com/paper/deep-learning-on-graphs-a-survey |
Repo | https://github.com/NorthPolesky/GNNpaper |
Framework | none |
Statistically Motivated Second Order Pooling
Title | Statistically Motivated Second Order Pooling |
Authors | Kaicheng Yu, Mathieu Salzmann |
Abstract | Second-order pooling, a.k.a.~bilinear pooling, has proven effective for deep learning based visual recognition. However, the resulting second-order networks yield a final representation that is orders of magnitude larger than that of standard, first-order ones, making them memory-intensive and cumbersome to deploy. Here, we introduce a general, parametric compression strategy that can produce more compact representations than existing compression techniques, yet outperform both compressed and uncompressed second-order models. Our approach is motivated by a statistical analysis of the network’s activations, relying on operations that lead to a Gaussian-distributed final representation, as inherently used by first-order deep networks. As evidenced by our experiments, this lets us outperform the state-of-the-art first-order and second-order models on several benchmark recognition datasets. |
Tasks | |
Published | 2018-01-23 |
URL | http://arxiv.org/abs/1801.07492v3 |
http://arxiv.org/pdf/1801.07492v3.pdf | |
PWC | https://paperswithcode.com/paper/statistically-motivated-second-order-pooling-1 |
Repo | https://github.com/kcyu2014/smsop |
Framework | tf |
Auto-Keras: An Efficient Neural Architecture Search System
Title | Auto-Keras: An Efficient Neural Architecture Search System |
Authors | Haifeng Jin, Qingquan Song, Xia Hu |
Abstract | Neural architecture search (NAS) has been proposed to automatically tune deep neural networks, but existing search algorithms, e.g., NASNet, PNAS, usually suffer from expensive computational cost. Network morphism, which keeps the functionality of a neural network while changing its neural architecture, could be helpful for NAS by enabling more efficient training during the search. In this paper, we propose a novel framework enabling Bayesian optimization to guide the network morphism for efficient neural architecture search. The framework develops a neural network kernel and a tree-structured acquisition function optimization algorithm to efficiently explores the search space. Intensive experiments on real-world benchmark datasets have been done to demonstrate the superior performance of the developed framework over the state-of-the-art methods. Moreover, we build an open-source AutoML system based on our method, namely Auto-Keras. The system runs in parallel on CPU and GPU, with an adaptive search strategy for different GPU memory limits. |
Tasks | AutoML, Neural Architecture Search |
Published | 2018-06-27 |
URL | http://arxiv.org/abs/1806.10282v3 |
http://arxiv.org/pdf/1806.10282v3.pdf | |
PWC | https://paperswithcode.com/paper/auto-keras-efficient-neural-architecture |
Repo | https://github.com/wpsliu123/AUTOKERAS |
Framework | none |
Information Aggregation via Dynamic Routing for Sequence Encoding
Title | Information Aggregation via Dynamic Routing for Sequence Encoding |
Authors | Jingjing Gong, Xipeng Qiu, Shaojing Wang, Xuanjing Huang |
Abstract | While much progress has been made in how to encode a text sequence into a sequence of vectors, less attention has been paid to how to aggregate these preceding vectors (outputs of RNN/CNN) into fixed-size encoding vector. Usually, a simple max or average pooling is used, which is a bottom-up and passive way of aggregation and lack of guidance by task information. In this paper, we propose an aggregation mechanism to obtain a fixed-size encoding with a dynamic routing policy. The dynamic routing policy is dynamically deciding that what and how much information need be transferred from each word to the final encoding of the text sequence. Following the work of Capsule Network, we design two dynamic routing policies to aggregate the outputs of RNN/CNN encoding layer into a final encoding vector. Compared to the other aggregation methods, dynamic routing can refine the messages according to the state of final encoding vector. Experimental results on five text classification tasks show that our method outperforms other aggregating models by a significant margin. Related source code is released on our github page. |
Tasks | Sentiment Analysis, Text Classification |
Published | 2018-06-05 |
URL | http://arxiv.org/abs/1806.01501v1 |
http://arxiv.org/pdf/1806.01501v1.pdf | |
PWC | https://paperswithcode.com/paper/information-aggregation-via-dynamic-routing |
Repo | https://github.com/FudanNLP/Capsule4TextClassification |
Framework | tf |
Don’t Use Large Mini-Batches, Use Local SGD
Title | Don’t Use Large Mini-Batches, Use Local SGD |
Authors | Tao Lin, Sebastian U. Stich, Kumar Kshitij Patel, Martin Jaggi |
Abstract | Mini-batch stochastic gradient methods (SGD) are state of the art for distributed training of deep neural networks. Drastic increases in the mini-batch sizes have lead to key efficiency and scalability gains in recent years. However, progress faces a major roadblock, as models trained with large batches often do not generalize well, i.e. they do not show good accuracy on new data. As a remedy, we propose a \emph{post-local} SGD and show that it significantly improves the generalization performance compared to large-batch training on standard benchmarks while enjoying the same efficiency (time-to-accuracy) and scalability. We further provide an extensive study of the communication efficiency vs. performance trade-offs associated with a host of \emph{local SGD} variants. |
Tasks | |
Published | 2018-08-22 |
URL | https://arxiv.org/abs/1808.07217v6 |
https://arxiv.org/pdf/1808.07217v6.pdf | |
PWC | https://paperswithcode.com/paper/dont-use-large-mini-batches-use-local-sgd |
Repo | https://github.com/BPrasad123/Data_Science_Learning |
Framework | tf |
General-to-Detailed GAN for Infrequent Class Medical Images
Title | General-to-Detailed GAN for Infrequent Class Medical Images |
Authors | Tatsuki Koga, Naoki Nonaka, Jun Sakuma, Jun Seita |
Abstract | Deep learning has significant potential for medical imaging. However, since the incident rate of each disease varies widely, the frequency of classes in a medical image dataset is imbalanced, leading to poor accuracy for such infrequent classes. One possible solution is data augmentation of infrequent classes using synthesized images created by Generative Adversarial Networks (GANs), but conventional GANs also require certain amount of images to learn. To overcome this limitation, here we propose General-to-detailed GAN (GDGAN), serially connected two GANs, one for general labels and the other for detailed labels. GDGAN produced diverse medical images, and the network trained with an augmented dataset outperformed other networks using existing methods with respect to Area-Under-Curve (AUC) of Receiver Operating Characteristic (ROC) curve. |
Tasks | Data Augmentation |
Published | 2018-11-28 |
URL | http://arxiv.org/abs/1812.01690v1 |
http://arxiv.org/pdf/1812.01690v1.pdf | |
PWC | https://paperswithcode.com/paper/general-to-detailed-gan-for-infrequent-class |
Repo | https://github.com/seitalab/GDGAN.pytorch |
Framework | pytorch |
Symbolic inductive bias for visually grounded learning of spoken language
Title | Symbolic inductive bias for visually grounded learning of spoken language |
Authors | Grzegorz Chrupała |
Abstract | A widespread approach to processing spoken language is to first automatically transcribe it into text. An alternative is to use an end-to-end approach: recent works have proposed to learn semantic embeddings of spoken language from images with spoken captions, without an intermediate transcription step. We propose to use multitask learning to exploit existing transcribed speech within the end-to-end setting. We describe a three-task architecture which combines the objectives of matching spoken captions with corresponding images, speech with text, and text with images. We show that the addition of the speech/text task leads to substantial performance improvements on image retrieval when compared to training the speech/image task in isolation. We conjecture that this is due to a strong inductive bias transcribed speech provides to the model, and offer supporting evidence for this. |
Tasks | Image Retrieval |
Published | 2018-12-21 |
URL | https://arxiv.org/abs/1812.09244v3 |
https://arxiv.org/pdf/1812.09244v3.pdf | |
PWC | https://paperswithcode.com/paper/symbolic-inductive-bias-for-visually-grounded |
Repo | https://github.com/gchrupala/symbolic-bias |
Framework | none |