Paper Group AWR 182
Calibrated Prediction Intervals for Neural Network Regressors. CompNet: Complementary Segmentation Network for Brain MRI Extraction. Geo-Supervised Visual Depth Prediction. Parameter Transfer Extreme Learning Machine based on Projective Model. Bio-inspired digit recognition using reward-modulated spike-timing-dependent plasticity in deep convolutio …
Calibrated Prediction Intervals for Neural Network Regressors
Title | Calibrated Prediction Intervals for Neural Network Regressors |
Authors | Gil Keren, Nicholas Cummins, Björn Schuller |
Abstract | Ongoing developments in neural network models are continually advancing the state of the art in terms of system accuracy. However, the predicted labels should not be regarded as the only core output; also important is a well-calibrated estimate of the prediction uncertainty. Such estimates and their calibration are critical in many practical applications. Despite their obvious aforementioned advantage in relation to accuracy, contemporary neural networks can, generally, be regarded as poorly calibrated and as such do not produce reliable output probability estimates. Further, while post-processing calibration solutions can be found in the relevant literature, these tend to be for systems performing classification. In this regard, we herein present two novel methods for acquiring calibrated predictions intervals for neural network regressors: empirical calibration and temperature scaling. In experiments using different regression tasks from the audio and computer vision domains, we find that both our proposed methods are indeed capable of producing calibrated prediction intervals for neural network regressors with any desired confidence level, a finding that is consistent across all datasets and neural network architectures we experimented with. In addition, we derive an additional practical recommendation for producing more accurate calibrated prediction intervals. We release the source code implementing our proposed methods for computing calibrated predicted intervals. The code for computing calibrated predicted intervals is publicly available. |
Tasks | Calibration |
Published | 2018-03-26 |
URL | http://arxiv.org/abs/1803.09546v3 |
http://arxiv.org/pdf/1803.09546v3.pdf | |
PWC | https://paperswithcode.com/paper/calibrated-prediction-intervals-for-neural |
Repo | https://github.com/cruvadom/Prediction_Intervals |
Framework | none |
CompNet: Complementary Segmentation Network for Brain MRI Extraction
Title | CompNet: Complementary Segmentation Network for Brain MRI Extraction |
Authors | Raunak Dey, Yi Hong |
Abstract | Brain extraction is a fundamental step for most brain imaging studies. In this paper, we investigate the problem of skull stripping and propose complementary segmentation networks (CompNets) to accurately extract the brain from T1-weighted MRI scans, for both normal and pathological brain images. The proposed networks are designed in the framework of encoder-decoder networks and have two pathways to learn features from both the brain tissue and its complementary part located outside of the brain. The complementary pathway extracts the features in the non-brain region and leads to a robust solution to brain extraction from MRIs with pathologies, which do not exist in our training dataset. We demonstrate the effectiveness of our networks by evaluating them on the OASIS dataset, resulting in the state of the art performance under the two-fold cross-validation setting. Moreover, the robustness of our networks is verified by testing on images with introduced pathologies and by showing its invariance to unseen brain pathologies. In addition, our complementary network design is general and can be extended to address other image segmentation problems with better generalization. |
Tasks | Semantic Segmentation, Skull Stripping |
Published | 2018-03-27 |
URL | http://arxiv.org/abs/1804.00521v2 |
http://arxiv.org/pdf/1804.00521v2.pdf | |
PWC | https://paperswithcode.com/paper/compnet-complementary-segmentation-network |
Repo | https://github.com/raun1/ISBI-2020-LITS_Hybrid_Comp_Net |
Framework | tf |
Geo-Supervised Visual Depth Prediction
Title | Geo-Supervised Visual Depth Prediction |
Authors | Xiaohan Fei, Alex Wong, Stefano Soatto |
Abstract | We propose using global orientation from inertial measurements, and the bias it induces on the shape of objects populating the scene, to inform visual 3D reconstruction. We test the effect of using the resulting prior in depth prediction from a single image, where the normal vectors to surfaces of objects of certain classes tend to align with gravity or be orthogonal to it. Adding such a prior to baseline methods for monocular depth prediction yields improvements beyond the state-of-the-art and illustrates the power of gravity as a supervisory signal. |
Tasks | 3D Reconstruction, Depth Estimation |
Published | 2018-07-30 |
URL | https://arxiv.org/abs/1807.11130v4 |
https://arxiv.org/pdf/1807.11130v4.pdf | |
PWC | https://paperswithcode.com/paper/geo-supervised-visual-depth-prediction |
Repo | https://github.com/feixh/GeoSup |
Framework | tf |
Parameter Transfer Extreme Learning Machine based on Projective Model
Title | Parameter Transfer Extreme Learning Machine based on Projective Model |
Authors | Chao Chen, Boyuan Jiang, Xinyu Jin |
Abstract | Recent years, transfer learning has attracted much attention in the community of machine learning. In this paper, we mainly focus on the tasks of parameter transfer under the framework of extreme learning machine (ELM). Unlike the existing parameter transfer approaches, which incorporate the source model information into the target by regularizing the di erence between the source and target domain parameters, an intuitively appealing projective-model is proposed to bridge the source and target model parameters. Specifically, we formulate the parameter transfer in the ELM networks by the means of parameter projection, and train the model by optimizing the projection matrix and classifier parameters jointly. Further more, the `L2,1-norm structured sparsity penalty is imposed on the source domain parameters, which encourages the joint feature selection and parameter transfer. To evaluate the e ectiveness of the proposed method, comprehensive experiments on several commonly used domain adaptation datasets are presented. The results show that the proposed method significantly outperforms the non-transfer ELM networks and other classical transfer learning methods. | |
Tasks | Domain Adaptation, Feature Selection, Transfer Learning |
Published | 2018-09-04 |
URL | http://arxiv.org/abs/1809.01018v2 |
http://arxiv.org/pdf/1809.01018v2.pdf | |
PWC | https://paperswithcode.com/paper/parameter-transfer-extreme-learning-machine |
Repo | https://github.com/BoyuanJiang/PTELM |
Framework | none |
Bio-inspired digit recognition using reward-modulated spike-timing-dependent plasticity in deep convolutional networks
Title | Bio-inspired digit recognition using reward-modulated spike-timing-dependent plasticity in deep convolutional networks |
Authors | Milad Mozafari, Mohammad Ganjtabesh, Abbas Nowzari-Dalini, Simon J. Thorpe, Timothée Masquelier |
Abstract | The primate visual system has inspired the development of deep artificial neural networks, which have revolutionized the computer vision domain. Yet these networks are much less energy-efficient than their biological counterparts, and they are typically trained with backpropagation, which is extremely data-hungry. To address these limitations, we used a deep convolutional spiking neural network (DCSNN) and a latency-coding scheme. We trained it using a combination of spike-timing-dependent plasticity (STDP) for the lower layers and reward-modulated STDP (R-STDP) for the higher ones. In short, with R-STDP a correct (resp. incorrect) decision leads to STDP (resp. anti-STDP). This approach led to an accuracy of $97.2%$ on MNIST, without requiring an external classifier. In addition, we demonstrated that R-STDP extracts features that are diagnostic for the task at hand, and discards the other ones, whereas STDP extracts any feature that repeats. Finally, our approach is biologically plausible, hardware friendly, and energy-efficient. |
Tasks | |
Published | 2018-03-31 |
URL | https://arxiv.org/abs/1804.00227v3 |
https://arxiv.org/pdf/1804.00227v3.pdf | |
PWC | https://paperswithcode.com/paper/bio-inspired-digit-recognition-using-spike |
Repo | https://github.com/miladmozafari/SpykeTorch |
Framework | pytorch |
ColNet: Embedding the Semantics of Web Tables for Column Type Prediction
Title | ColNet: Embedding the Semantics of Web Tables for Column Type Prediction |
Authors | Jiaoyan Chen, Ernesto Jimenez-Ruiz, Ian Horrocks, Charles Sutton |
Abstract | Automatically annotating column types with knowledge base (KB) concepts is a critical task to gain a basic understanding of web tables. Current methods rely on either table metadata like column name or entity correspondences of cells in the KB, and may fail to deal with growing web tables with incomplete meta information. In this paper we propose a neural network based column type annotation framework named ColNet which is able to integrate KB reasoning and lookup with machine learning and can automatically train Convolutional Neural Networks for prediction. The prediction model not only considers the contextual semantics within a cell using word representation, but also embeds the semantics of a column by learning locality features from multiple cells. The method is evaluated with DBPedia and two different web table datasets, T2Dv2 from the general Web and Limaye from Wikipedia pages, and achieves higher performance than the state-of-the-art approaches. |
Tasks | |
Published | 2018-11-04 |
URL | http://arxiv.org/abs/1811.01304v2 |
http://arxiv.org/pdf/1811.01304v2.pdf | |
PWC | https://paperswithcode.com/paper/colnet-embedding-the-semantics-of-web-tables |
Repo | https://github.com/alan-turing-institute/SemAIDA |
Framework | none |
Classifying and Visualizing Emotions with Emotional DAN
Title | Classifying and Visualizing Emotions with Emotional DAN |
Authors | Ivona Tautkute, Tomasz Trzcinski |
Abstract | Classification of human emotions remains an important and challenging task for many computer vision algorithms, especially in the era of humanoid robots which coexist with humans in their everyday life. Currently proposed methods for emotion recognition solve this task using multi-layered convolutional networks that do not explicitly infer any facial features in the classification phase. In this work, we postulate a fundamentally different approach to solve emotion recognition task that relies on incorporating facial landmarks as a part of the classification loss function. To that end, we extend a recently proposed Deep Alignment Network (DAN) with a term related to facial features. Thanks to this simple modification, our model called EmotionalDAN is able to outperform state-of-the-art emotion classification methods on two challenging benchmark dataset by up to 5%. Furthermore, we visualize image regions analyzed by the network when making a decision and the results indicate that our EmotionalDAN model is able to correctly identify facial landmarks responsible for expressing the emotions. |
Tasks | Emotion Classification, Emotion Recognition |
Published | 2018-10-23 |
URL | http://arxiv.org/abs/1810.10529v1 |
http://arxiv.org/pdf/1810.10529v1.pdf | |
PWC | https://paperswithcode.com/paper/classifying-and-visualizing-emotions-with |
Repo | https://github.com/IvonaTau/emotionaldan |
Framework | tf |
Combatting Adversarial Attacks through Denoising and Dimensionality Reduction: A Cascaded Autoencoder Approach
Title | Combatting Adversarial Attacks through Denoising and Dimensionality Reduction: A Cascaded Autoencoder Approach |
Authors | Rajeev Sahay, Rehana Mahfuz, Aly El Gamal |
Abstract | Machine Learning models are vulnerable to adversarial attacks that rely on perturbing the input data. This work proposes a novel strategy using Autoencoder Deep Neural Networks to defend a machine learning model against two gradient-based attacks: The Fast Gradient Sign attack and Fast Gradient attack. First we use an autoencoder to denoise the test data, which is trained with both clean and corrupted data. Then, we reduce the dimension of the denoised data using the hidden layer representation of another autoencoder. We perform this experiment for multiple values of the bound of adversarial perturbations, and consider different numbers of reduced dimensions. When the test data is preprocessed using this cascaded pipeline, the tested deep neural network classifier yields a much higher accuracy, thus mitigating the effect of the adversarial perturbation. |
Tasks | Denoising, Dimensionality Reduction |
Published | 2018-12-07 |
URL | http://arxiv.org/abs/1812.03087v1 |
http://arxiv.org/pdf/1812.03087v1.pdf | |
PWC | https://paperswithcode.com/paper/combatting-adversarial-attacks-through |
Repo | https://github.com/rajeevsahay/ae-defenses |
Framework | none |
The Price of Fair PCA: One Extra Dimension
Title | The Price of Fair PCA: One Extra Dimension |
Authors | Samira Samadi, Uthaipon Tantipongpipat, Jamie Morgenstern, Mohit Singh, Santosh Vempala |
Abstract | We investigate whether the standard dimensionality reduction technique of PCA inadvertently produces data representations with different fidelity for two different populations. We show on several real-world data sets, PCA has higher reconstruction error on population A than on B (for example, women versus men or lower- versus higher-educated individuals). This can happen even when the data set has a similar number of samples from A and B. This motivates our study of dimensionality reduction techniques which maintain similar fidelity for A and B. We define the notion of Fair PCA and give a polynomial-time algorithm for finding a low dimensional representation of the data which is nearly-optimal with respect to this measure. Finally, we show on real-world data sets that our algorithm can be used to efficiently generate a fair low dimensional representation of the data. |
Tasks | Dimensionality Reduction |
Published | 2018-10-31 |
URL | http://arxiv.org/abs/1811.00103v1 |
http://arxiv.org/pdf/1811.00103v1.pdf | |
PWC | https://paperswithcode.com/paper/the-price-of-fair-pca-one-extra-dimension |
Repo | https://github.com/samirasamadi/Fair-PCA |
Framework | none |
Timeception for Complex Action Recognition
Title | Timeception for Complex Action Recognition |
Authors | Noureldien Hussein, Efstratios Gavves, Arnold W. M. Smeulders |
Abstract | This paper focuses on the temporal aspect for recognizing human activities in videos; an important visual cue that has long been undervalued. We revisit the conventional definition of activity and restrict it to Complex Action: a set of one-actions with a weak temporal pattern that serves a specific purpose. Related works use spatiotemporal 3D convolutions with fixed kernel size, too rigid to capture the varieties in temporal extents of complex actions, and too short for long-range temporal modeling. In contrast, we use multi-scale temporal convolutions, and we reduce the complexity of 3D convolutions. The outcome is Timeception convolution layers, which reasons about minute-long temporal patterns, a factor of 8 longer than best related works. As a result, Timeception achieves impressive accuracy in recognizing the human activities of Charades, Breakfast Actions, and MultiTHUMOS. Further, we demonstrate that Timeception learns long-range temporal dependencies and tolerate temporal extents of complex actions. |
Tasks | Action Recognition In Videos |
Published | 2018-12-04 |
URL | http://arxiv.org/abs/1812.01289v2 |
http://arxiv.org/pdf/1812.01289v2.pdf | |
PWC | https://paperswithcode.com/paper/timeception-for-complex-action-recognition |
Repo | https://github.com/noureldien/timeception |
Framework | pytorch |
Learning with privileged information via adversarial discriminative modality distillation
Title | Learning with privileged information via adversarial discriminative modality distillation |
Authors | Nuno C. Garcia, Pietro Morerio, Vittorio Murino |
Abstract | Heterogeneous data modalities can provide complementary cues for several tasks, usually leading to more robust algorithms and better performance. However, while training data can be accurately collected to include a variety of sensory modalities, it is often the case that not all of them are available in real life (testing) scenarios, where a model has to be deployed. This raises the challenge of how to extract information from multimodal data in the training stage, in a form that can be exploited at test time, considering limitations such as noisy or missing modalities. This paper presents a new approach in this direction for RGB-D vision tasks, developed within the adversarial learning and privileged information frameworks. We consider the practical case of learning representations from depth and RGB videos, while relying only on RGB data at test time. We propose a new approach to train a hallucination network that learns to distill depth information via adversarial learning, resulting in a clean approach without several losses to balance or hyperparameters. We report state-of-the-art results on object classification on the NYUD dataset and video action recognition on the largest multimodal dataset available for this task, the NTU RGB+D, as well as on the Northwestern-UCLA. |
Tasks | Action Recognition In Videos, Object Classification |
Published | 2018-10-19 |
URL | https://arxiv.org/abs/1810.08437v2 |
https://arxiv.org/pdf/1810.08437v2.pdf | |
PWC | https://paperswithcode.com/paper/learning-with-privileged-information-via |
Repo | https://github.com/pmorerio/admd |
Framework | tf |
Extracting and Analyzing Semantic Relatedness between Cities Using News Articles
Title | Extracting and Analyzing Semantic Relatedness between Cities Using News Articles |
Authors | Yingjie Hu, Xinyue Ye, Shih-Lung Shaw |
Abstract | News articles capture a variety of topics about our society. They reflect not only the socioeconomic activities that happened in our physical world, but also some of the cultures, human interests, and public concerns that exist only in the perceptions of people. Cities are frequently mentioned in news articles, and two or more cities may co-occur in the same article. Such co-occurrence often suggests certain relatedness between the mentioned cities, and the relatedness may be under different topics depending on the contents of the news articles. We consider the relatedness under different topics as semantic relatedness. By reading news articles, one can grasp the general semantic relatedness between cities, yet, given hundreds of thousands of news articles, it is very difficult, if not impossible, for anyone to manually read them. This paper proposes a computational framework which can “read” a large number of news articles and extract the semantic relatedness between cities. This framework is based on a natural language processing model and employs a machine learning process to identify the main topics of news articles. We describe the overall structure of this framework and its individual modules, and then apply it to an experimental dataset with more than 500,000 news articles covering the top 100 U.S. cities spanning a 10-year period. We perform exploratory visualization of the extracted semantic relatedness under different topics and over multiple years. We also analyze the impact of geographic distance on semantic relatedness and find varied distance decay effects. The proposed framework can be used to support large-scale content analysis in city network research. |
Tasks | |
Published | 2018-09-08 |
URL | http://arxiv.org/abs/1809.02823v1 |
http://arxiv.org/pdf/1809.02823v1.pdf | |
PWC | https://paperswithcode.com/paper/extracting-and-analyzing-semantic-relatedness |
Repo | https://github.com/YingjieHu/CityRelatednessViaNews |
Framework | none |
Classification using margin pursuit
Title | Classification using margin pursuit |
Authors | Matthew J. Holland |
Abstract | In this work, we study a new approach to optimizing the margin distribution realized by binary classifiers. The classical approach to this problem is simply maximization of the expected margin, while more recent proposals consider simultaneous variance control and proxy objectives based on robust location estimates, in the vein of keeping the margin distribution sharply concentrated in a desirable region. While conceptually appealing, these new approaches are often computationally unwieldy, and theoretical guarantees are limited. Given this context, we propose an algorithm which searches the hypothesis space in such a way that a pre-set “margin level” ends up being a distribution-robust estimator of the margin location. This procedure is easily implemented using gradient descent, and admits finite-sample bounds on the excess risk under unbounded inputs. Empirical tests on real-world benchmark data reinforce the basic principles highlighted by the theory, and are suggestive of a promising new technique for classification. |
Tasks | |
Published | 2018-10-11 |
URL | http://arxiv.org/abs/1810.04863v1 |
http://arxiv.org/pdf/1810.04863v1.pdf | |
PWC | https://paperswithcode.com/paper/classification-using-margin-pursuit |
Repo | https://github.com/feedbackward/catcube |
Framework | none |
Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting
Title | Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting |
Authors | Yen-Chun Chen, Mohit Bansal |
Abstract | Inspired by how humans summarize long documents, we propose an accurate and fast summarization model that first selects salient sentences and then rewrites them abstractively (i.e., compresses and paraphrases) to generate a concise overall summary. We use a novel sentence-level policy gradient method to bridge the non-differentiable computation between these two neural networks in a hierarchical way, while maintaining language fluency. Empirically, we achieve the new state-of-the-art on all metrics (including human evaluation) on the CNN/Daily Mail dataset, as well as significantly higher abstractiveness scores. Moreover, by first operating at the sentence-level and then the word-level, we enable parallel decoding of our neural generative model that results in substantially faster (10-20x) inference speed as well as 4x faster training convergence than previous long-paragraph encoder-decoder models. We also demonstrate the generalization of our model on the test-only DUC-2002 dataset, where we achieve higher scores than a state-of-the-art model. |
Tasks | Abstractive Text Summarization |
Published | 2018-05-28 |
URL | http://arxiv.org/abs/1805.11080v1 |
http://arxiv.org/pdf/1805.11080v1.pdf | |
PWC | https://paperswithcode.com/paper/fast-abstractive-summarization-with-reinforce |
Repo | https://github.com/ChenRocks/fast_abs_rl |
Framework | pytorch |
Translating a Math Word Problem to an Expression Tree
Title | Translating a Math Word Problem to an Expression Tree |
Authors | Lei Wang, Yan Wang, Deng Cai, Dongxiang Zhang, Xiaojiang Liu |
Abstract | Sequence-to-sequence (SEQ2SEQ) models have been successfully applied to automatic math word problem solving. Despite its simplicity, a drawback still remains: a math word problem can be correctly solved by more than one equations. This non-deterministic transduction harms the performance of maximum likelihood estimation. In this paper, by considering the uniqueness of expression tree, we propose an equation normalization method to normalize the duplicated equations. Moreover, we analyze the performance of three popular SEQ2SEQ models on the math word problem solving. We find that each model has its own specialty in solving problems, consequently an ensemble model is then proposed to combine their advantages. Experiments on dataset Math23K show that the ensemble model with equation normalization significantly outperforms the previous state-of-the-art methods. |
Tasks | Math Word Problem Solving |
Published | 2018-11-14 |
URL | http://arxiv.org/abs/1811.05632v2 |
http://arxiv.org/pdf/1811.05632v2.pdf | |
PWC | https://paperswithcode.com/paper/translating-a-math-word-problem-to-an |
Repo | https://github.com/SumbeeLei/Math_EN |
Framework | pytorch |