Paper Group AWR 72
Topic Sensitive Attention on Generic Corpora Corrects Sense Bias in Pretrained Embeddings. RandAugment: Practical automated data augmentation with a reduced search space. TactileGCN: A Graph Convolutional Network for Predicting Grasp Stability with Tactile Sensors. Big Transfer (BiT): General Visual Representation Learning. Semantic Segmentation of …
Topic Sensitive Attention on Generic Corpora Corrects Sense Bias in Pretrained Embeddings
Title | Topic Sensitive Attention on Generic Corpora Corrects Sense Bias in Pretrained Embeddings |
Authors | Vihari Piratla, Sunita Sarawagi, Soumen Chakrabarti |
Abstract | Given a small corpus $\mathcal D_T$ pertaining to a limited set of focused topics, our goal is to train embeddings that accurately capture the sense of words in the topic in spite of the limited size of $\mathcal D_T$. These embeddings may be used in various tasks involving $\mathcal D_T$. A popular strategy in limited data settings is to adapt pre-trained embeddings $\mathcal E$ trained on a large corpus. To correct for sense drift, fine-tuning, regularization, projection, and pivoting have been proposed recently. Among these, regularization informed by a word’s corpus frequency performed well, but we improve upon it using a new regularizer based on the stability of its cooccurrence with other words. However, a thorough comparison across ten topics, spanning three tasks, with standardized settings of hyper-parameters, reveals that even the best embedding adaptation strategies provide small gains beyond well-tuned baselines, which many earlier comparisons ignored. In a bold departure from adapting pretrained embeddings, we propose using $\mathcal D_T$ to probe, attend to, and borrow fragments from any large, topic-rich source corpus (such as Wikipedia), which need not be the corpus used to pretrain embeddings. This step is made scalable and practical by suitable indexing. We reach the surprising conclusion that even limited corpus augmentation is more useful than adapting embeddings, which suggests that non-dominant sense information may be irrevocably obliterated from pretrained embeddings and cannot be salvaged by adaptation. |
Tasks | |
Published | 2019-06-05 |
URL | https://arxiv.org/abs/1906.02688v2 |
https://arxiv.org/pdf/1906.02688v2.pdf | |
PWC | https://paperswithcode.com/paper/topic-sensitive-attention-on-generic-corpora |
Repo | https://github.com/vihari/focussed_embs |
Framework | tf |
RandAugment: Practical automated data augmentation with a reduced search space
Title | RandAugment: Practical automated data augmentation with a reduced search space |
Authors | Ekin D. Cubuk, Barret Zoph, Jonathon Shlens, Quoc V. Le |
Abstract | Recent work has shown that data augmentation has the potential to significantly improve the generalization of deep learning models. Recently, automated augmentation strategies have led to state-of-the-art results in image classification and object detection. While these strategies were optimized for improving validation accuracy, they also led to state-of-the-art results in semi-supervised learning and improved robustness to common corruptions of images. An obstacle to a large-scale adoption of these methods is a separate search phase which increases the training complexity and may substantially increase the computational cost. Additionally, due to the separate search phase, these approaches are unable to adjust the regularization strength based on model or dataset size. Automated augmentation policies are often found by training small models on small datasets and subsequently applied to train larger models. In this work, we remove both of these obstacles. RandAugment has a significantly reduced search space which allows it to be trained on the target task with no need for a separate proxy task. Furthermore, due to the parameterization, the regularization strength may be tailored to different model and dataset sizes. RandAugment can be used uniformly across different tasks and datasets and works out of the box, matching or surpassing all previous automated augmentation approaches on CIFAR-10/100, SVHN, and ImageNet. On the ImageNet dataset we achieve 85.0% accuracy, a 0.6% increase over the previous state-of-the-art and 1.0% increase over baseline augmentation. On object detection, RandAugment leads to 1.0-1.3% improvement over baseline augmentation, and is within 0.3% mAP of AutoAugment on COCO. Finally, due to its interpretable hyperparameter, RandAugment may be used to investigate the role of data augmentation with varying model and dataset size. Code is available online. |
Tasks | Data Augmentation, Image Classification, Object Detection |
Published | 2019-09-30 |
URL | https://arxiv.org/abs/1909.13719v2 |
https://arxiv.org/pdf/1909.13719v2.pdf | |
PWC | https://paperswithcode.com/paper/randaugment-practical-data-augmentation-with |
Repo | https://github.com/etetteh/sota-data-augmentation-and-optimizers |
Framework | pytorch |
TactileGCN: A Graph Convolutional Network for Predicting Grasp Stability with Tactile Sensors
Title | TactileGCN: A Graph Convolutional Network for Predicting Grasp Stability with Tactile Sensors |
Authors | Alberto Garcia-Garcia, Brayan Stiven Zapata-Impata, Sergio Orts-Escolano, Pablo Gil, Jose Garcia-Rodriguez |
Abstract | Tactile sensors provide useful contact data during the interaction with an object which can be used to accurately learn to determine the stability of a grasp. Most of the works in the literature represented tactile readings as plain feature vectors or matrix-like tactile images, using them to train machine learning models. In this work, we explore an alternative way of exploiting tactile information to predict grasp stability by leveraging graph-like representations of tactile data, which preserve the actual spatial arrangement of the sensor’s taxels and their locality. In experimentation, we trained a Graph Neural Network to binary classify grasps as stable or slippery ones. To train such network and prove its predictive capabilities for the problem at hand, we captured a novel dataset of approximately 5000 three-fingered grasps across 41 objects for training and 1000 grasps with 10 unknown objects for testing. Our experiments prove that this novel approach can be effectively used to predict grasp stability. |
Tasks | |
Published | 2019-01-18 |
URL | http://arxiv.org/abs/1901.06181v1 |
http://arxiv.org/pdf/1901.06181v1.pdf | |
PWC | https://paperswithcode.com/paper/tactilegcn-a-graph-convolutional-network-for |
Repo | https://github.com/3dperceptionlab/biotacsp-stability-set-v2 |
Framework | none |
Big Transfer (BiT): General Visual Representation Learning
Title | Big Transfer (BiT): General Visual Representation Learning |
Authors | Alexander Kolesnikov, Lucas Beyer, Xiaohua Zhai, Joan Puigcerver, Jessica Yung, Sylvain Gelly, Neil Houlsby |
Abstract | Transfer of pre-trained representations improves sample efficiency and simplifies hyperparameter tuning when training deep neural networks for vision. We revisit the paradigm of pre-training on large supervised datasets and fine-tuning the model on a target task. We scale up pre-training, and propose a simple recipe that we call Big Transfer (BiT). By combining a few carefully selected components, and transferring using a simple heuristic, we achieve strong performance on over 20 datasets. BiT performs well across a surprisingly wide range of data regimes – from 1 example per class to 1M total examples. BiT achieves 87.5% top-1 accuracy on ILSVRC-2012, 99.4% on CIFAR-10, and 76.3% on the 19 task Visual Task Adaptation Benchmark (VTAB). On small datasets, BiT attains 76.8% on ILSVRC-2012 with 10 examples per class, and 97.0% on CIFAR-10 with 10 examples per class. We conduct detailed analysis of the main components that lead to high transfer performance. |
Tasks | Fine-Grained Image Classification, Image Classification, Representation Learning |
Published | 2019-12-24 |
URL | https://arxiv.org/abs/1912.11370v2 |
https://arxiv.org/pdf/1912.11370v2.pdf | |
PWC | https://paperswithcode.com/paper/large-scale-learning-of-general-visual |
Repo | https://github.com/SoojungYang/supervised_pretraining_GN_WS |
Framework | tf |
Semantic Segmentation of Panoramic Images Using a Synthetic Dataset
Title | Semantic Segmentation of Panoramic Images Using a Synthetic Dataset |
Authors | Yuanyou Xu, Kaiwei Wang, Kailun Yang, Dongming Sun, Jia Fu |
Abstract | Panoramic images have advantages in information capacity and scene stability due to their large field of view (FoV). In this paper, we propose a method to synthesize a new dataset of panoramic image. We managed to stitch the images taken from different directions into panoramic images, together with their labeled images, to yield the panoramic semantic segmentation dataset denominated as SYNTHIA-PANO. For the purpose of finding out the effect of using panoramic images as training dataset, we designed and performed a comprehensive set of experiments. Experimental results show that using panoramic images as training data is beneficial to the segmentation result. In addition, it has been shown that by using panoramic images with a 180 degree FoV as training data the model has better performance. Furthermore, the model trained with panoramic images also has a better capacity to resist the image distortion. |
Tasks | Semantic Segmentation |
Published | 2019-09-02 |
URL | https://arxiv.org/abs/1909.00532v1 |
https://arxiv.org/pdf/1909.00532v1.pdf | |
PWC | https://paperswithcode.com/paper/semantic-segmentation-of-panoramic-images |
Repo | https://github.com/Francis515/SYNTHIA-PANO |
Framework | none |
Spatial and Colour Opponency in Anatomically Constrained Deep Networks
Title | Spatial and Colour Opponency in Anatomically Constrained Deep Networks |
Authors | Ethan Harris, Daniela Mihai, Jonathon Hare |
Abstract | Colour vision has long fascinated scientists, who have sought to understand both the physiology of the mechanics of colour vision and the psychophysics of colour perception. We consider representations of colour in anatomically constrained convolutional deep neural networks. Following ideas from neuroscience, we classify cells in early layers into groups relating to their spectral and spatial functionality. We show the emergence of single and double opponent cells in our networks and characterise how the distribution of these cells changes under the constraint of a retinal bottleneck. Our experiments not only open up a new understanding of how deep networks process spatial and colour information, but also provide new tools to help understand the black box of deep learning. The code for all experiments is avaialable at \url{https://github.com/ecs-vlc/opponency}. |
Tasks | |
Published | 2019-10-14 |
URL | https://arxiv.org/abs/1910.11086v1 |
https://arxiv.org/pdf/1910.11086v1.pdf | |
PWC | https://paperswithcode.com/paper/spatial-and-colour-opponency-in-anatomically |
Repo | https://github.com/ecs-vlc/opponency |
Framework | pytorch |
AtomNAS: Fine-Grained End-to-End Neural Architecture Search
Title | AtomNAS: Fine-Grained End-to-End Neural Architecture Search |
Authors | Jieru Mei, Yingwei Li, Xiaochen Lian, Xiaojie Jin, Linjie Yang, Alan Yuille, Jianchao Yang |
Abstract | Search space design is very critical to neural architecture search (NAS) algorithms. We propose a fine-grained search space comprised of atomic blocks, a minimal search unit that is much smaller than the ones used in recent NAS algorithms. This search space allows a mix of operations by composing different types of atomic blocks, while the search space in previous methods only allows homogeneous operations. Based on this search space, we propose a resource-aware architecture search framework which automatically assigns the computational resources (e.g., output channel numbers) for each operation by jointly considering the performance and the computational cost. In addition, to accelerate the search process, we propose a dynamic network shrinkage technique which prunes the atomic blocks with negligible influence on outputs on the fly. Instead of a search-and-retrain two-stage paradigm, our method simultaneously searches and trains the target architecture. Our method achieves state-of-the-art performance under several FLOPs configurations on ImageNet with a small searching cost. We open our entire codebase at: https://github.com/meijieru/AtomNAS. |
Tasks | Neural Architecture Search |
Published | 2019-12-20 |
URL | https://arxiv.org/abs/1912.09640v2 |
https://arxiv.org/pdf/1912.09640v2.pdf | |
PWC | https://paperswithcode.com/paper/atomnas-fine-grained-end-to-end-neural-1 |
Repo | https://github.com/meijieru/AtomNAS |
Framework | pytorch |
An α-Matte Boundary Defocus Model Based Cascaded Network for Multi-focus Image Fusion
Title | An α-Matte Boundary Defocus Model Based Cascaded Network for Multi-focus Image Fusion |
Authors | Haoyu Ma, Qingmin Liao, Juncheng Zhang, Shaojun Liu, Jing-Hao Xue |
Abstract | Capturing an all-in-focus image with a single camera is difficult since the depth of field of the camera is usually limited. An alternative method to obtain the all-in-focus image is to fuse several images focusing at different depths. However, existing multi-focus image fusion methods cannot obtain clear results for areas near the focused/defocused boundary (FDB). In this paper, a novel {\alpha}-matte boundary defocus model is proposed to generate realistic training data with the defocus spread effect precisely modeled, especially for areas near the FDB. Based on this {\alpha}-matte defocus model and the generated data, a cascaded boundary aware convolutional network termed MMF-Net is proposed and trained, aiming to achieve clearer fusion results around the FDB. More specifically, the MMF-Net consists of two cascaded sub-nets for initial fusion and boundary fusion, respectively; these two sub-nets are designed to first obtain a guidance map of FDB and then refine the fusion near the FDB. Experiments demonstrate that with the help of the new {\alpha}-matte boundary defocus model, the proposed MMF-Net outperforms the state-of-the-art methods both qualitatively and quantitatively. |
Tasks | |
Published | 2019-10-29 |
URL | https://arxiv.org/abs/1910.13136v2 |
https://arxiv.org/pdf/1910.13136v2.pdf | |
PWC | https://paperswithcode.com/paper/an-matte-boundary-defocus-model-based |
Repo | https://github.com/xytmhy/MMF-Net-alpha-Matte-Boundary-Defocus-Model-Fusion |
Framework | pytorch |
Ablation Studies in Artificial Neural Networks
Title | Ablation Studies in Artificial Neural Networks |
Authors | Richard Meyes, Melanie Lu, Constantin Waubert de Puiseau, Tobias Meisen |
Abstract | Ablation studies have been widely used in the field of neuroscience to tackle complex biological systems such as the extensively studied Drosophila central nervous system, the vertebrate brain and more interestingly and most delicately, the human brain. In the past, these kinds of studies were utilized to uncover structure and organization in the brain, i.e. a mapping of features inherent to external stimuli onto different areas of the neocortex. considering the growth in size and complexity of state-of-the-art artificial neural networks (ANNs) and the corresponding growth in complexity of the tasks that are tackled by these networks, the question arises whether ablation studies may be used to investigate these networks for a similar organization of their inner representations. In this paper, we address this question and performed two ablation studies in two fundamentally different ANNs to investigate their inner representations of two well-known benchmark datasets from the computer vision domain. We found that features distinct to the local and global structure of the data are selectively represented in specific parts of the network. Furthermore, some of these representations are redundant, awarding the network a certain robustness to structural damages. We further determined the importance of specific parts of the network for the classification task solely based on the weight structure of single units. Finally, we examined the ability of damaged networks to recover from the consequences of ablations by means of recovery training. We argue that ablations studies are a feasible method to investigate knowledge representations in ANNs and are especially helpful to examine a networks robustness to structural damages, a feature of ANNs that will become increasingly important for future safety-critical applications. |
Tasks | |
Published | 2019-01-24 |
URL | http://arxiv.org/abs/1901.08644v2 |
http://arxiv.org/pdf/1901.08644v2.pdf | |
PWC | https://paperswithcode.com/paper/ablation-studies-in-artificial-neural |
Repo | https://github.com/RichardMeyes/AblationStudies |
Framework | pytorch |
Conditional Density Estimation Tools in Python and R with Applications to Photometric Redshifts and Likelihood-Free Cosmological Inference
Title | Conditional Density Estimation Tools in Python and R with Applications to Photometric Redshifts and Likelihood-Free Cosmological Inference |
Authors | Niccolò Dalmasso, Taylor Pospisil, Ann B. Lee, Rafael Izbicki, Peter E. Freeman, Alex I. Malz |
Abstract | It is well known in astronomy that propagating non-Gaussian prediction uncertainty in photometric redshift estimates is key to reducing bias in downstream cosmological analyses. Similarly, likelihood-free inference approaches, which are beginning to emerge as a tool for cosmological analysis, require a characterization of the full uncertainty landscape of the parameters of interest given observed data. However, most machine learning (ML) or training-based methods with open-source software target point prediction or classification, and hence fall short in quantifying uncertainty in complex regression and parameter inference settings. As an alternative to methods that focus on predicting the response (or parameters) $\mathbf{y}$ from features $\mathbf{x}$, we provide nonparametric conditional density estimation (CDE) tools for approximating and validating the entire probability density function (PDF) $\mathrm{p}(\mathbf{y}\mathbf{x})$ of $\mathbf{y}$ given (i.e., conditional on) $\mathbf{x}$. As there is no one-size-fits-all CDE method, the goal of this work is to provide a comprehensive range of statistical tools and open-source software for nonparametric CDE and method assessment which can accommodate different types of settings and be easily fit to the problem at hand. Specifically, we introduce four CDE software packages in $\texttt{Python}$ and $\texttt{R}$ based on ML prediction methods adapted and optimized for CDE: $\texttt{NNKCDE}$, $\texttt{RFCDE}$, $\texttt{FlexCode}$, and $\texttt{DeepCDE}$. Furthermore, we present the $\texttt{cdetools}$ package, which includes functions for computing a CDE loss function for tuning and assessing the quality of individual PDFs, along with diagnostic functions. We provide sample code in $\texttt{Python}$ and $\texttt{R}$ as well as examples of applications to photometric redshift estimation and likelihood-free cosmological inference via CDE. |
Tasks | Density Estimation, Model Selection, Photometric Redshift Estimation |
Published | 2019-08-30 |
URL | https://arxiv.org/abs/1908.11523v2 |
https://arxiv.org/pdf/1908.11523v2.pdf | |
PWC | https://paperswithcode.com/paper/conditional-density-estimation-tools-in |
Repo | https://github.com/Mr8ND/cdetools_applications |
Framework | none |
Purifying Interaction Effects with the Functional ANOVA: An Efficient Algorithm for Recovering Identifiable Additive Models
Title | Purifying Interaction Effects with the Functional ANOVA: An Efficient Algorithm for Recovering Identifiable Additive Models |
Authors | Benjamin Lengerich, Sarah Tan, Chun-Hao Chang, Giles Hooker, Rich Caruana |
Abstract | Recent methods for training generalized additive models (GAMs) with pairwise interactions achieve state-of-the-art accuracy on a variety of datasets. Adding interactions to GAMs, however, introduces an identifiability problem: effects can be freely moved between main effects and interaction effects without changing the model predictions. In some cases, this can lead to contradictory interpretations of the same underlying function. This is a critical problem because a central motivation of GAMs is model interpretability. In this paper, we use the Functional ANOVA decomposition to uniquely define interaction effects and thus produce identifiable additive models with purified interactions. To compute this decomposition, we present a fast, exact, mass-moving algorithm that transforms any piecewise-constant function (such as a tree-based model) into a purified, canonical representation. We apply this algorithm to several datasets and show large disparity, including contradictions, between the apparent and the purified effects. |
Tasks | |
Published | 2019-11-12 |
URL | https://arxiv.org/abs/1911.04974v1 |
https://arxiv.org/pdf/1911.04974v1.pdf | |
PWC | https://paperswithcode.com/paper/purifying-interaction-effects-with-the |
Repo | https://github.com/microsoft/interpret |
Framework | none |
Hierarchical Gating Networks for Sequential Recommendation
Title | Hierarchical Gating Networks for Sequential Recommendation |
Authors | Chen Ma, Peng Kang, Xue Liu |
Abstract | The chronological order of user-item interactions is a key feature in many recommender systems, where the items that users will interact may largely depend on those items that users just accessed recently. However, with the tremendous increase of users and items, sequential recommender systems still face several challenging problems: (1) the hardness of modeling the long-term user interests from sparse implicit feedback; (2) the difficulty of capturing the short-term user interests given several items the user just accessed. To cope with these challenges, we propose a hierarchical gating network (HGN), integrated with the Bayesian Personalized Ranking (BPR) to capture both the long-term and short-term user interests. Our HGN consists of a feature gating module, an instance gating module, and an item-item product module. In particular, our feature gating and instance gating modules select what item features can be passed to the downstream layers from the feature and instance levels, respectively. Our item-item product module explicitly captures the item relations between the items that users accessed in the past and those items users will access in the future. We extensively evaluate our model with several state-of-the-art methods and different validation metrics on five real-world datasets. The experimental results demonstrate the effectiveness of our model on Top-N sequential recommendation. |
Tasks | Recommendation Systems |
Published | 2019-06-21 |
URL | https://arxiv.org/abs/1906.09217v1 |
https://arxiv.org/pdf/1906.09217v1.pdf | |
PWC | https://paperswithcode.com/paper/hierarchical-gating-networks-for-sequential |
Repo | https://github.com/wuliwei9278/HGN_baseline |
Framework | pytorch |
Partial Scanning Transmission Electron Microscopy with Deep Learning
Title | Partial Scanning Transmission Electron Microscopy with Deep Learning |
Authors | Jeffrey M. Ede, Richard Beanland |
Abstract | Compressed sensing algorithms are used to decrease electron microscope scan time and electron beam exposure with minimal information loss. Following successful applications of deep learning to compressed sensing, we have developed a two-stage multiscale generative adversarial neural network to complete realistic 512$\times$512 scanning transmission electron micrographs from spiral, jittered gridlike, and other partial scans. For spiral scans and mean squared error based pre-training, this enables electron beam coverage to be decreased by 17.9$\times$ with a 3.8% test set root mean squared intensity error, and by 87.0$\times$ with a 6.2% error. Our generator networks are trained on partial scans created from a new dataset of 16227 scanning transmission electron micrographs. High performance is achieved with adaptive learning rate clipping of loss spikes and an auxiliary trainer network. Our source code, new dataset, and pre-trained models have been made publicly available at https://github.com/Jeffrey-Ede/partial-STEM |
Tasks | |
Published | 2019-05-31 |
URL | https://arxiv.org/abs/1905.13667v2 |
https://arxiv.org/pdf/1905.13667v2.pdf | |
PWC | https://paperswithcode.com/paper/partial-scan-electron-microscopy-with-deep |
Repo | https://github.com/Jeffrey-Ede/ALRC |
Framework | tf |
Alzheimer’s Disease Brain MRI Classification: Challenges and Insights
Title | Alzheimer’s Disease Brain MRI Classification: Challenges and Insights |
Authors | Yi Ren Fung, Ziqiang Guan, Ritesh Kumar, Joie Yeahuay Wu, Madalina Fiterau |
Abstract | In recent years, many papers have reported state-of-the-art performance on Alzheimer’s Disease classification with MRI scans from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) dataset using convolutional neural networks. However, we discover that when we split that data into training and testing sets at the subject level, we are not able to obtain similar performance, bringing the validity of many of the previous studies into question. Furthermore, we point out that previous works use different subsets of the ADNI data, making comparison across similar works tricky. In this study, we present the results of three splitting methods, discuss the motivations behind their validity, and report our results using all of the available subjects. |
Tasks | |
Published | 2019-06-10 |
URL | https://arxiv.org/abs/1906.04231v1 |
https://arxiv.org/pdf/1906.04231v1.pdf | |
PWC | https://paperswithcode.com/paper/alzheimers-disease-brain-mri-classification |
Repo | https://github.com/Information-Fusion-Lab-Umass/alzheimers-cnn-study |
Framework | pytorch |
Boundary-Aware Network for Fast and High-Accuracy Portrait Segmentation
Title | Boundary-Aware Network for Fast and High-Accuracy Portrait Segmentation |
Authors | Xi Chen, Donglian Qi, Jianxin Shen |
Abstract | Compared with other semantic segmentation tasks, portrait segmentation requires both higher precision and faster inference speed. However, this problem has not been well studied in previous works. In this paper, we propose a lightweight network architecture, called Boundary-Aware Network (BANet) which selectively extracts detail information in boundary area to make high-quality segmentation output with real-time( >25FPS) speed. In addition, we design a new loss function called refine loss which supervises the network with image level gradient information. Our model is able to produce finer segmentation results which has richer details than annotations. |
Tasks | Semantic Segmentation |
Published | 2019-01-12 |
URL | http://arxiv.org/abs/1901.03814v1 |
http://arxiv.org/pdf/1901.03814v1.pdf | |
PWC | https://paperswithcode.com/paper/boundary-aware-network-for-fast-and-high |
Repo | https://github.com/lewisluk/BoundaryAwareNetwork |
Framework | tf |