Paper Group ANR 873
Investigating the Impact of CNN Depth on Neonatal Seizure Detection Performance. Change Point Methods on a Sequence of Graphs. Elements of Effective Deep Reinforcement Learning towards Tactical Driving Decision Making. DeepPap: Deep Convolutional Networks for Cervical Cell Classification. Towards Single-phase Single-stage Detection of Pulmonary Nod …
Investigating the Impact of CNN Depth on Neonatal Seizure Detection Performance
Title | Investigating the Impact of CNN Depth on Neonatal Seizure Detection Performance |
Authors | Alison O’Shea, Gordon Lightbody, Geraldine Boylan, Andriy Temko |
Abstract | This study presents a novel, deep, fully convolutional architecture which is optimized for the task of EEG-based neonatal seizure detection. Architectures of different depths were designed and tested; varying network depth impacts convolutional receptive fields and the corresponding learned feature complexity. Two deep convolutional networks are compared with a shallow SVM-based neonatal seizure detector, which relies on the extraction of hand-crafted features. On a large clinical dataset, of over 800 hours of multichannel unedited EEG, containing 1389 seizure events, the deep 11-layer architecture significantly outperforms the shallower architectures, improving the AUC90 from 82.6% to 86.8%. Combining the end-to-end deep architecture with the feature-based shallow SVM further improves the AUC90 to 87.6%. The fusion of classifiers of different depths gives greatly improved performance and reduced variability, making the combined classifier more clinically reliable. |
Tasks | EEG, Seizure Detection |
Published | 2018-06-08 |
URL | http://arxiv.org/abs/1806.03044v1 |
http://arxiv.org/pdf/1806.03044v1.pdf | |
PWC | https://paperswithcode.com/paper/investigating-the-impact-of-cnn-depth-on |
Repo | |
Framework | |
Change Point Methods on a Sequence of Graphs
Title | Change Point Methods on a Sequence of Graphs |
Authors | Daniele Zambon, Cesare Alippi, Lorenzo Livi |
Abstract | Given a finite sequence of graphs, e.g., coming from technological, biological, and social networks, the paper proposes a methodology to identify possible changes in stationarity in the stochastic process generating the graphs. In order to cover a large class of applications, we consider the general family of attributed graphs where both topology (number of vertexes and edge configuration) and related attributes are allowed to change also in the stationary case. Novel Change Point Methods (CPMs) are proposed, that (i) map graphs into a vector domain; (ii) apply a suitable statistical test in the vector space; (iii) detect the change –if any– according to a confidence level and provide an estimate for its time occurrence. Two specific multivariate CPMs have been designed: one that detects shifts in the distribution mean, the other addressing generic changes affecting the distribution. We ground our proposal with theoretical results showing how to relate the inference attained in the numerical vector space to the graph domain, and vice versa. We also show how to extend the methodology for handling multiple change points in the same sequence. Finally, the proposed CPMs have been validated on real data sets coming from epileptic-seizure detection problems and on labeled data sets for graph classification. Results show the effectiveness of what proposed in relevant application scenarios. |
Tasks | Graph Classification, Seizure Detection |
Published | 2018-05-18 |
URL | http://arxiv.org/abs/1805.07113v2 |
http://arxiv.org/pdf/1805.07113v2.pdf | |
PWC | https://paperswithcode.com/paper/change-point-methods-on-a-sequence-of-graphs |
Repo | |
Framework | |
Elements of Effective Deep Reinforcement Learning towards Tactical Driving Decision Making
Title | Elements of Effective Deep Reinforcement Learning towards Tactical Driving Decision Making |
Authors | Jingchu Liu, Pengfei Hou, Lisen Mu, Yinan Yu, Chang Huang |
Abstract | Tactical driving decision making is crucial for autonomous driving systems and has attracted considerable interest in recent years. In this paper, we propose several practical components that can speed up deep reinforcement learning algorithms towards tactical decision making tasks: 1) non-uniform action skipping as a more stable alternative to action-repetition frame skipping, 2) a counter-based penalty for lanes on which ego vehicle has less right-of-road, and 3) heuristic inference-time action masking for apparently undesirable actions. We evaluate the proposed components in a realistic driving simulator and compare them with several baselines. Results show that the proposed scheme provides superior performance in terms of safety, efficiency, and comfort. |
Tasks | Autonomous Driving, Decision Making |
Published | 2018-02-01 |
URL | http://arxiv.org/abs/1802.00332v1 |
http://arxiv.org/pdf/1802.00332v1.pdf | |
PWC | https://paperswithcode.com/paper/elements-of-effective-deep-reinforcement |
Repo | |
Framework | |
DeepPap: Deep Convolutional Networks for Cervical Cell Classification
Title | DeepPap: Deep Convolutional Networks for Cervical Cell Classification |
Authors | Ling Zhang, Le Lu, Isabella Nogues, Ronald M. Summers, Shaoxiong Liu, Jianhua Yao |
Abstract | Automation-assisted cervical screening via Pap smear or liquid-based cytology (LBC) is a highly effective cell imaging based cancer detection tool, where cells are partitioned into “abnormal” and “normal” categories. However, the success of most traditional classification methods relies on the presence of accurate cell segmentations. Despite sixty years of research in this field, accurate segmentation remains a challenge in the presence of cell clusters and pathologies. Moreover, previous classification methods are only built upon the extraction of hand-crafted features, such as morphology and texture. This paper addresses these limitations by proposing a method to directly classify cervical cells - without prior segmentation - based on deep features, using convolutional neural networks (ConvNets). First, the ConvNet is pre-trained on a natural image dataset. It is subsequently fine-tuned on a cervical cell dataset consisting of adaptively re-sampled image patches coarsely centered on the nuclei. In the testing phase, aggregation is used to average the prediction scores of a similar set of image patches. The proposed method is evaluated on both Pap smear and LBC datasets. Results show that our method outperforms previous algorithms in classification accuracy (98.3%), area under the curve (AUC) (0.99) values, and especially specificity (98.3%), when applied to the Herlev benchmark Pap smear dataset and evaluated using five-fold cross-validation. Similar superior performances are also achieved on the HEMLBC (H&E stained manual LBC) dataset. Our method is promising for the development of automation-assisted reading systems in primary cervical screening. |
Tasks | |
Published | 2018-01-25 |
URL | http://arxiv.org/abs/1801.08616v1 |
http://arxiv.org/pdf/1801.08616v1.pdf | |
PWC | https://paperswithcode.com/paper/deeppap-deep-convolutional-networks-for |
Repo | |
Framework | |
Towards Single-phase Single-stage Detection of Pulmonary Nodules in Chest CT Imaging
Title | Towards Single-phase Single-stage Detection of Pulmonary Nodules in Chest CT Imaging |
Authors | Zhongliu Xie |
Abstract | Detection of pulmonary nodules in chest CT imaging plays a crucial role in early diagnosis of lung cancer. Manual examination is highly time-consuming and error prone, calling for computer-aided detection, both to improve efficiency and reduce misdiagnosis. Over the years, a range of systems have been proposed, mostly following a two-phase paradigm with: 1) candidate detection, 2) false positive reduction. Recently, deep learning has become a dominant force in algorithm development. As for candidate detection, prior art was mainly based on the two-stage Faster R-CNN framework, which starts with an initial sub-net to generate a set of class-agnostic region proposals, followed by a second sub-net to perform classification and bounding-box regression. In contrast, we abandon the conventional two-phase paradigm and two-stage framework altogether and propose to train a single network for end-to-end nodule detection instead, without transfer learning or further post-processing. Our feature learning model is a modification of the ResNet and feature pyramid network combined, powered by RReLU activation. The major challenge is the condition of extreme inter-class and intra-class sample imbalance, where the positives are overwhelmed by a large negative pool, which is mostly composed of easy and a handful of hard negatives. Direct training on all samples can seriously undermine training efficacy. We propose a patch-based sampling strategy over a set of regularly updating anchors, which narrows sampling scope to all positives and only hard negatives, effectively addressing this issue. As a result, our approach substantially outperforms prior art in terms of both accuracy and speed. Finally, the prevailing FROC evaluation over [1/8, 1/4, 1/2, 1, 2, 4, 8] false positives per scan, is far from ideal in real clinical environments. We suggest FROC over [1, 2, 4] false positives as a better metric. |
Tasks | Transfer Learning |
Published | 2018-07-16 |
URL | http://arxiv.org/abs/1807.05972v1 |
http://arxiv.org/pdf/1807.05972v1.pdf | |
PWC | https://paperswithcode.com/paper/towards-single-phase-single-stage-detection |
Repo | |
Framework | |
A Robust AUC Maximization Framework with Simultaneous Outlier Detection and Feature Selection for Positive-Unlabeled Classification
Title | A Robust AUC Maximization Framework with Simultaneous Outlier Detection and Feature Selection for Positive-Unlabeled Classification |
Authors | Ke Ren, Haichuan Yang, Yu Zhao, Mingshan Xue, Hongyu Miao, Shuai Huang, Ji Liu |
Abstract | The positive-unlabeled (PU) classification is a common scenario in real-world applications such as healthcare, text classification, and bioinformatics, in which we only observe a few samples labeled as “positive” together with a large volume of “unlabeled” samples that may contain both positive and negative samples. Building robust classifier for the PU problem is very challenging, especially for complex data where the negative samples overwhelm and mislabeled samples or corrupted features exist. To address these three issues, we propose a robust learning framework that unifies AUC maximization (a robust metric for biased labels), outlier detection (for excluding wrong labels), and feature selection (for excluding corrupted features). The generalization error bounds are provided for the proposed model that give valuable insight into the theoretical performance of the method and lead to useful practical guidance, e.g., to train a model, we find that the included unlabeled samples are sufficient as long as the sample size is comparable to the number of positive samples in the training process. Empirical comparisons and two real-world applications on surgical site infection (SSI) and EEG seizure detection are also conducted to show the effectiveness of the proposed model. |
Tasks | EEG, Feature Selection, Outlier Detection, Seizure Detection, Text Classification |
Published | 2018-03-18 |
URL | http://arxiv.org/abs/1803.06604v1 |
http://arxiv.org/pdf/1803.06604v1.pdf | |
PWC | https://paperswithcode.com/paper/a-robust-auc-maximization-framework-with |
Repo | |
Framework | |
JointDNN: An Efficient Training and Inference Engine for Intelligent Mobile Cloud Computing Services
Title | JointDNN: An Efficient Training and Inference Engine for Intelligent Mobile Cloud Computing Services |
Authors | Amir Erfan Eshratifar, Mohammad Saeed Abrishami, Massoud Pedram |
Abstract | Deep learning models are being deployed in many mobile intelligent applications. End-side services, such as intelligent personal assistants, autonomous cars, and smart home services often employ either simple local models on the mobile or complex remote models on the cloud. However, recent studies have shown that partitioning the DNN computations between the mobile and cloud can increase the latency and energy efficiencies. In this paper, we propose an efficient, adaptive, and practical engine, JointDNN, for collaborative computation between a mobile device and cloud for DNNs in both inference and training phase. JointDNN not only provides an energy and performance efficient method of querying DNNs for the mobile side but also benefits the cloud server by reducing the amount of its workload and communications compared to the cloud-only approach. Given the DNN architecture, we investigate the efficiency of processing some layers on the mobile device and some layers on the cloud server. We provide optimization formulations at layer granularity for forward- and backward-propagations in DNNs, which can adapt to mobile battery limitations and cloud server load constraints and quality of service. JointDNN achieves up to 18 and 32 times reductions on the latency and mobile energy consumption of querying DNNs compared to the status-quo approaches, respectively. |
Tasks | |
Published | 2018-01-25 |
URL | https://arxiv.org/abs/1801.08618v2 |
https://arxiv.org/pdf/1801.08618v2.pdf | |
PWC | https://paperswithcode.com/paper/jointdnn-an-efficient-training-and-inference |
Repo | |
Framework | |
Predicting the Effects of News Sentiments on the Stock Market
Title | Predicting the Effects of News Sentiments on the Stock Market |
Authors | Dev Shah, Haruna Isah, Farhana Zulkernine |
Abstract | Stock market forecasting is very important in the planning of business activities. Stock price prediction has attracted many researchers in multiple disciplines including computer science, statistics, economics, finance, and operations research. Recent studies have shown that the vast amount of online information in the public domain such as Wikipedia usage pattern, news stories from the mainstream media, and social media discussions can have an observable effect on investors opinions towards financial markets. The reliability of the computational models on stock market prediction is important as it is very sensitive to the economy and can directly lead to financial loss. In this paper, we retrieved, extracted, and analyzed the effects of news sentiments on the stock market. Our main contributions include the development of a sentiment analysis dictionary for the financial sector, the development of a dictionary-based sentiment analysis model, and the evaluation of the model for gauging the effects of news sentiments on stocks for the pharmaceutical market. Using only news sentiments, we achieved a directional accuracy of 70.59% in predicting the trends in short-term stock price movement. |
Tasks | Sentiment Analysis, Stock Market Prediction, Stock Price Prediction |
Published | 2018-12-11 |
URL | http://arxiv.org/abs/1812.04199v1 |
http://arxiv.org/pdf/1812.04199v1.pdf | |
PWC | https://paperswithcode.com/paper/predicting-the-effects-of-news-sentiments-on |
Repo | |
Framework | |
Deep Classification of Epileptic Signals
Title | Deep Classification of Epileptic Signals |
Authors | David Ahmedt-Aristizabal, Clinton Fookes, Kien Nguyen, Sridha Sridharan |
Abstract | Electrophysiological observation plays a major role in epilepsy evaluation. However, human interpretation of brain signals is subjective and prone to misdiagnosis. Automating this process, especially seizure detection relying on scalp-based Electroencephalography (EEG) and intracranial EEG, has been the focus of research over recent decades. Nevertheless, its numerous challenges have inhibited a definitive solution. Inspired by recent advances in deep learning, we propose a new classification approach for EEG time series based on Recurrent Neural Networks (RNNs) via the use of Long-Short Term Memory (LSTM) networks. The proposed deep network effectively learns and models discriminative temporal patterns from EEG sequential data. Especially, the features are automatically discovered from the raw EEG data without any pre-processing step, eliminating humans from laborious feature design task. We also show that, in the epilepsy scenario, simple architectures can achieve competitive performance. Using simple architectures significantly benefits in the practical scenario considering their low computation complexity and reduced requirement for large training datasets. Using a public dataset, a multi-fold cross-validation scheme exhibited an average validation accuracy of 95.54% and an average AUC of 0.9582 of the ROC curve among all sets defined in the experiment. This work reinforces the benefits of deep learning to be further attended in clinical applications and neuroscientific research. |
Tasks | EEG, Seizure Detection, Time Series |
Published | 2018-01-11 |
URL | http://arxiv.org/abs/1801.03610v1 |
http://arxiv.org/pdf/1801.03610v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-classification-of-epileptic-signals |
Repo | |
Framework | |
Label Sanitization against Label Flipping Poisoning Attacks
Title | Label Sanitization against Label Flipping Poisoning Attacks |
Authors | Andrea Paudice, Luis Muñoz-González, Emil C. Lupu |
Abstract | Many machine learning systems rely on data collected in the wild from untrusted sources, exposing the learning algorithms to data poisoning. Attackers can inject malicious data in the training dataset to subvert the learning process, compromising the performance of the algorithm producing errors in a targeted or an indiscriminate way. Label flipping attacks are a special case of data poisoning, where the attacker can control the labels assigned to a fraction of the training points. Even if the capabilities of the attacker are constrained, these attacks have been shown to be effective to significantly degrade the performance of the system. In this paper we propose an efficient algorithm to perform optimal label flipping poisoning attacks and a mechanism to detect and relabel suspicious data points, mitigating the effect of such poisoning attacks. |
Tasks | data poisoning |
Published | 2018-03-02 |
URL | http://arxiv.org/abs/1803.00992v2 |
http://arxiv.org/pdf/1803.00992v2.pdf | |
PWC | https://paperswithcode.com/paper/label-sanitization-against-label-flipping |
Repo | |
Framework | |
Learning Incremental Triplet Margin for Person Re-identification
Title | Learning Incremental Triplet Margin for Person Re-identification |
Authors | Yingying Zhang, Qiaoyong Zhong, Liang Ma, Di Xie, Shiliang Pu |
Abstract | Person re-identification (ReID) aims to match people across multiple non-overlapping video cameras deployed at different locations. To address this challenging problem, many metric learning approaches have been proposed, among which triplet loss is one of the state-of-the-arts. In this work, we explore the margin between positive and negative pairs of triplets and prove that large margin is beneficial. In particular, we propose a novel multi-stage training strategy which learns incremental triplet margin and improves triplet loss effectively. Multiple levels of feature maps are exploited to make the learned features more discriminative. Besides, we introduce global hard identity searching method to sample hard identities when generating a training batch. Extensive experiments on Market-1501, CUHK03, and DukeMTMCreID show that our approach yields a performance boost and outperforms most existing state-of-the-art methods. |
Tasks | Metric Learning, Person Re-Identification |
Published | 2018-12-17 |
URL | http://arxiv.org/abs/1812.06576v1 |
http://arxiv.org/pdf/1812.06576v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-incremental-triplet-margin-for |
Repo | |
Framework | |
Kalman filter demystified: from intuition to probabilistic graphical model to real case in financial markets
Title | Kalman filter demystified: from intuition to probabilistic graphical model to real case in financial markets |
Authors | Eric Benhamou |
Abstract | In this paper, we revisit the Kalman filter theory. After giving the intuition on a simplified financial markets example, we revisit the maths underlying it. We then show that Kalman filter can be presented in a very different fashion using graphical models. This enables us to establish the connection between Kalman filter and Hidden Markov Models. We then look at their application in financial markets and provide various intuitions in terms of their applicability for complex systems such as financial markets. Although this paper has been written more like a self contained work connecting Kalman filter to Hidden Markov Models and hence revisiting well known and establish results, it contains new results and brings additional contributions to the field. First, leveraging on the link between Kalman filter and HMM, it gives new algorithms for inference for extended Kalman filters. Second, it presents an alternative to the traditional estimation of parameters using EM algorithm thanks to the usage of CMA-ES optimization. Third, it examines the application of Kalman filter and its Hidden Markov models version to financial markets, providing various dynamics assumptions and tests. We conclude by connecting Kalman filter approach to trend following technical analysis system and showing their superior performances for trend following detection. |
Tasks | |
Published | 2018-11-28 |
URL | http://arxiv.org/abs/1811.11618v2 |
http://arxiv.org/pdf/1811.11618v2.pdf | |
PWC | https://paperswithcode.com/paper/kalman-filter-demystified-from-intuition-to |
Repo | |
Framework | |
Label Denoising with Large Ensembles of Heterogeneous Neural Networks
Title | Label Denoising with Large Ensembles of Heterogeneous Neural Networks |
Authors | Pavel Ostyakov, Elizaveta Logacheva, Roman Suvorov, Vladimir Aliev, Gleb Sterkin, Oleg Khomenko, Sergey I. Nikolenko |
Abstract | Despite recent advances in computer vision based on various convolutional architectures, video understanding remains an important challenge. In this work, we present and discuss a top solution for the large-scale video classification (labeling) problem introduced as a Kaggle competition based on the YouTube-8M dataset. We show and compare different approaches to preprocessing, data augmentation, model architectures, and model combination. Our final model is based on a large ensemble of video- and frame-level models but fits into rather limiting hardware constraints. We apply an approach based on knowledge distillation to deal with noisy labels in the original dataset and the recently developed mixup technique to improve the basic models. |
Tasks | Data Augmentation, Denoising, Video Classification, Video Understanding |
Published | 2018-09-12 |
URL | http://arxiv.org/abs/1809.04403v2 |
http://arxiv.org/pdf/1809.04403v2.pdf | |
PWC | https://paperswithcode.com/paper/label-denoising-with-large-ensembles-of |
Repo | |
Framework | |
Adversarial adaptive 1-D convolutional neural networks for bearing fault diagnosis under varying working condition
Title | Adversarial adaptive 1-D convolutional neural networks for bearing fault diagnosis under varying working condition |
Authors | Bo Zhang, Wei Li, Jie Hao, Xiao-Li Li, Meng Zhang |
Abstract | Traditional intelligent fault diagnosis of rolling bearings work well only under a common assumption that the labeled training data (source domain) and unlabeled testing data (target domain) are drawn from the same distribution. However, in many real-world applications, this assumption does not hold, especially when the working condition varies. In this paper, a new adversarial adaptive 1-D CNN called A2CNN is proposed to address this problem. A2CNN consists of four parts, namely, a source feature extractor, a target feature extractor, a label classifier and a domain discriminator. The layers between the source and target feature extractor are partially untied during the training stage to take both training efficiency and domain adaptation into consideration. Experiments show that A2CNN has strong fault-discriminative and domain-invariant capacity, and therefore can achieve high accuracy under different working conditions. We also visualize the learned features and the networks to explore the reasons behind the high performance of our proposed model. |
Tasks | Domain Adaptation |
Published | 2018-05-01 |
URL | http://arxiv.org/abs/1805.00778v3 |
http://arxiv.org/pdf/1805.00778v3.pdf | |
PWC | https://paperswithcode.com/paper/adversarial-adaptive-1-d-convolutional-neural |
Repo | |
Framework | |
Dimensionality Reduction in Deep Learning for Chest X-Ray Analysis of Lung Cancer
Title | Dimensionality Reduction in Deep Learning for Chest X-Ray Analysis of Lung Cancer |
Authors | Yu. Gordienko, Yu. Kochura, O. Alienin, O. Rokovyi, S. Stirenko, Peng Gang, Jiang Hui, Wei Zeng |
Abstract | Efficiency of some dimensionality reduction techniques, like lung segmentation, bone shadow exclusion, and t-distributed stochastic neighbor embedding (t-SNE) for exclusion of outliers, is estimated for analysis of chest X-ray (CXR) 2D images by deep learning approach to help radiologists identify marks of lung cancer in CXR. Training and validation of the simple convolutional neural network (CNN) was performed on the open JSRT dataset (dataset #01), the JSRT after bone shadow exclusion - BSE-JSRT (dataset #02), JSRT after lung segmentation (dataset #03), BSE-JSRT after lung segmentation (dataset #04), and segmented BSE-JSRT after exclusion of outliers by t-SNE method (dataset #05). The results demonstrate that the pre-processed dataset obtained after lung segmentation, bone shadow exclusion, and filtering out the outliers by t-SNE (dataset #05) demonstrates the highest training rate and best accuracy in comparison to the other pre-processed datasets. |
Tasks | Dimensionality Reduction |
Published | 2018-01-19 |
URL | http://arxiv.org/abs/1801.06495v1 |
http://arxiv.org/pdf/1801.06495v1.pdf | |
PWC | https://paperswithcode.com/paper/dimensionality-reduction-in-deep-learning-for |
Repo | |
Framework | |