Paper Group AWR 109
Semi-Analytic Resampling in Lasso. Referring Relationships. Text normalization using memory augmented neural networks. ESPNet: Efficient Spatial Pyramid of Dilated Convolutions for Semantic Segmentation. A Recurrent Graph Neural Network for Multi-Relational Data. Modular meta-learning. Recent Advances in Object Detection in the Age of Deep Convolut …
Semi-Analytic Resampling in Lasso
Title | Semi-Analytic Resampling in Lasso |
Authors | Tomoyuki Obuchi, Yoshiyuki Kabashima |
Abstract | An approximate method for conducting resampling in Lasso, the $\ell_1$ penalized linear regression, in a semi-analytic manner is developed, whereby the average over the resampled datasets is directly computed without repeated numerical sampling, thus enabling an inference free of the statistical fluctuations due to sampling finiteness, as well as a significant reduction of computational time. The proposed method is based on a message passing type algorithm, and its fast convergence is guaranteed by the state evolution analysis, when covariates are provided as zero-mean independently and identically distributed Gaussian random variables. It is employed to implement bootstrapped Lasso (Bolasso) and stability selection, both of which are variable selection methods using resampling in conjunction with Lasso, and resolves their disadvantage regarding computational cost. To examine approximation accuracy and efficiency, numerical experiments were carried out using simulated datasets. Moreover, an application to a real-world dataset, the wine quality dataset, is presented. To process such real-world datasets, an objective criterion for determining the relevance of selected variables is also introduced by the addition of noise variables and resampling. |
Tasks | |
Published | 2018-02-28 |
URL | http://arxiv.org/abs/1802.10254v2 |
http://arxiv.org/pdf/1802.10254v2.pdf | |
PWC | https://paperswithcode.com/paper/semi-analytic-resampling-in-lasso |
Repo | https://github.com/T-Obuchi/AMPR_lasso_matlab |
Framework | none |
Referring Relationships
Title | Referring Relationships |
Authors | Ranjay Krishna, Ines Chami, Michael Bernstein, Li Fei-Fei |
Abstract | Images are not simply sets of objects: each image represents a web of interconnected relationships. These relationships between entities carry semantic meaning and help a viewer differentiate between instances of an entity. For example, in an image of a soccer match, there may be multiple persons present, but each participates in different relationships: one is kicking the ball, and the other is guarding the goal. In this paper, we formulate the task of utilizing these “referring relationships” to disambiguate between entities of the same category. We introduce an iterative model that localizes the two entities in the referring relationship, conditioned on one another. We formulate the cyclic condition between the entities in a relationship by modelling predicates that connect the entities as shifts in attention from one entity to another. We demonstrate that our model can not only outperform existing approaches on three datasets — CLEVR, VRD and Visual Genome — but also that it produces visually meaningful predicate shifts, as an instance of interpretable neural networks. Finally, we show that by modelling predicates as attention shifts, we can even localize entities in the absence of their category, allowing our model to find completely unseen categories. |
Tasks | |
Published | 2018-03-28 |
URL | http://arxiv.org/abs/1803.10362v2 |
http://arxiv.org/pdf/1803.10362v2.pdf | |
PWC | https://paperswithcode.com/paper/referring-relationships |
Repo | https://github.com/shikorab/DSG |
Framework | tf |
Text normalization using memory augmented neural networks
Title | Text normalization using memory augmented neural networks |
Authors | Subhojeet Pramanik, Aman Hussain |
Abstract | We perform text normalization, i.e. the transformation of words from the written to the spoken form, using a memory augmented neural network. With the addition of dynamic memory access and storage mechanism, we present a neural architecture that will serve as a language-agnostic text normalization system while avoiding the kind of unacceptable errors made by the LSTM-based recurrent neural networks. By successfully reducing the frequency of such mistakes, we show that this novel architecture is indeed a better alternative. Our proposed system requires significantly lesser amounts of data, training time and compute resources. Additionally, we perform data up-sampling, circumventing the data sparsity problem in some semiotic classes, to show that sufficient examples in any particular class can improve the performance of our text normalization system. Although a few occurrences of these errors still remain in certain semiotic classes, we demonstrate that memory augmented networks with meta-learning capabilities can open many doors to a superior text normalization system. |
Tasks | Meta-Learning |
Published | 2018-05-31 |
URL | http://arxiv.org/abs/1806.00044v3 |
http://arxiv.org/pdf/1806.00044v3.pdf | |
PWC | https://paperswithcode.com/paper/text-normalization-using-memory-augmented |
Repo | https://github.com/cognibit/Text-Normalization-Demo |
Framework | tf |
ESPNet: Efficient Spatial Pyramid of Dilated Convolutions for Semantic Segmentation
Title | ESPNet: Efficient Spatial Pyramid of Dilated Convolutions for Semantic Segmentation |
Authors | Sachin Mehta, Mohammad Rastegari, Anat Caspi, Linda Shapiro, Hannaneh Hajishirzi |
Abstract | We introduce a fast and efficient convolutional neural network, ESPNet, for semantic segmentation of high resolution images under resource constraints. ESPNet is based on a new convolutional module, efficient spatial pyramid (ESP), which is efficient in terms of computation, memory, and power. ESPNet is 22 times faster (on a standard GPU) and 180 times smaller than the state-of-the-art semantic segmentation network PSPNet, while its category-wise accuracy is only 8% less. We evaluated ESPNet on a variety of semantic segmentation datasets including Cityscapes, PASCAL VOC, and a breast biopsy whole slide image dataset. Under the same constraints on memory and computation, ESPNet outperforms all the current efficient CNN networks such as MobileNet, ShuffleNet, and ENet on both standard metrics and our newly introduced performance metrics that measure efficiency on edge devices. Our network can process high resolution images at a rate of 112 and 9 frames per second on a standard GPU and edge device, respectively. |
Tasks | Real-Time Semantic Segmentation, Semantic Segmentation |
Published | 2018-03-19 |
URL | http://arxiv.org/abs/1803.06815v3 |
http://arxiv.org/pdf/1803.06815v3.pdf | |
PWC | https://paperswithcode.com/paper/espnet-efficient-spatial-pyramid-of-dilated |
Repo | https://github.com/adichaloo/EdgeNet |
Framework | pytorch |
A Recurrent Graph Neural Network for Multi-Relational Data
Title | A Recurrent Graph Neural Network for Multi-Relational Data |
Authors | Vassilis N. Ioannidis, Antonio G. Marques, Georgios B. Giannakis |
Abstract | The era of data deluge has sparked the interest in graph-based learning methods in a number of disciplines such as sociology, biology, neuroscience, or engineering. In this paper, we introduce a graph recurrent neural network (GRNN) for scalable semi-supervised learning from multi-relational data. Key aspects of the novel GRNN architecture are the use of multi-relational graphs, the dynamic adaptation to the different relations via learnable weights, and the consideration of graph-based regularizers to promote smoothness and alleviate over-parametrization. Our ultimate goal is to design a powerful learning architecture able to: discover complex and highly non-linear data associations, combine (and select) multiple types of relations, and scale gracefully with respect to the size of the graph. Numerical tests with real data sets corroborate the design goals and illustrate the performance gains relative to competing alternatives. |
Tasks | |
Published | 2018-11-05 |
URL | http://arxiv.org/abs/1811.02061v3 |
http://arxiv.org/pdf/1811.02061v3.pdf | |
PWC | https://paperswithcode.com/paper/a-recurrent-graph-neural-network-for-multi |
Repo | https://github.com/bioannidis/adaptive_recurrent_graph_neural_network |
Framework | tf |
Modular meta-learning
Title | Modular meta-learning |
Authors | Ferran Alet, Tomás Lozano-Pérez, Leslie P. Kaelbling |
Abstract | Many prediction problems, such as those that arise in the context of robotics, have a simplifying underlying structure that, if known, could accelerate learning. In this paper, we present a strategy for learning a set of neural network modules that can be combined in different ways. We train different modular structures on a set of related tasks and generalize to new tasks by composing the learned modules in new ways. By reusing modules to generalize we achieve combinatorial generalization, akin to the “infinite use of finite means” displayed in language. Finally, we show this improves performance in two robotics-related problems. |
Tasks | Meta-Learning |
Published | 2018-06-26 |
URL | https://arxiv.org/abs/1806.10166v2 |
https://arxiv.org/pdf/1806.10166v2.pdf | |
PWC | https://paperswithcode.com/paper/modular-meta-learning |
Repo | https://github.com/FerranAlet/modular-metalearning |
Framework | pytorch |
Recent Advances in Object Detection in the Age of Deep Convolutional Neural Networks
Title | Recent Advances in Object Detection in the Age of Deep Convolutional Neural Networks |
Authors | Shivang Agarwal, Jean Ogier Du Terrail, Frédéric Jurie |
Abstract | Object detection-the computer vision task dealing with detecting instances of objects of a certain class (e.g., ‘car’, ‘plane’, etc.) in images-attracted a lot of attention from the community during the last 5 years. This strong interest can be explained not only by the importance this task has for many applications but also by the phenomenal advances in this area since the arrival of deep convolutional neural networks (DCNN). This article reviews the recent literature on object detection with deep CNN, in a comprehensive way, and provides an in-depth view of these recent advances. The survey covers not only the typical architectures (SSD, YOLO, Faster-RCNN) but also discusses the challenges currently met by the community and goes on to show how the problem of object detection can be extended. This survey also reviews the public datasets and associated state-of-the-art algorithms. |
Tasks | Object Detection |
Published | 2018-09-10 |
URL | https://arxiv.org/abs/1809.03193v2 |
https://arxiv.org/pdf/1809.03193v2.pdf | |
PWC | https://paperswithcode.com/paper/recent-advances-in-object-detection-in-the |
Repo | https://github.com/atheheath/papers |
Framework | none |
Conditional BERT Contextual Augmentation
Title | Conditional BERT Contextual Augmentation |
Authors | Xing Wu, Shangwen Lv, Liangjun Zang, Jizhong Han, Songlin Hu |
Abstract | We propose a novel data augmentation method for labeled sentences called conditional BERT contextual augmentation. Data augmentation methods are often applied to prevent overfitting and improve generalization of deep neural network models. Recently proposed contextual augmentation augments labeled sentences by randomly replacing words with more varied substitutions predicted by language model. BERT demonstrates that a deep bidirectional language model is more powerful than either an unidirectional language model or the shallow concatenation of a forward and backward model. We retrofit BERT to conditional BERT by introducing a new conditional masked language model\footnote{The term “conditional masked language model” appeared once in original BERT paper, which indicates context-conditional, is equivalent to term “masked language model”. In our paper, “conditional masked language model” indicates we apply extra label-conditional constraint to the “masked language model”.} task. The well trained conditional BERT can be applied to enhance contextual augmentation. Experiments on six various different text classification tasks show that our method can be easily applied to both convolutional or recurrent neural networks classifier to obtain obvious improvement. |
Tasks | Data Augmentation, Language Modelling, Text Classification |
Published | 2018-12-17 |
URL | http://arxiv.org/abs/1812.06705v1 |
http://arxiv.org/pdf/1812.06705v1.pdf | |
PWC | https://paperswithcode.com/paper/conditional-bert-contextual-augmentation |
Repo | https://github.com/Gal1eo/DD2424 |
Framework | pytorch |
Generalizable Adversarial Training via Spectral Normalization
Title | Generalizable Adversarial Training via Spectral Normalization |
Authors | Farzan Farnia, Jesse M. Zhang, David Tse |
Abstract | Deep neural networks (DNNs) have set benchmarks on a wide array of supervised learning tasks. Trained DNNs, however, often lack robustness to minor adversarial perturbations to the input, which undermines their true practicality. Recent works have increased the robustness of DNNs by fitting networks using adversarially-perturbed training samples, but the improved performance can still be far below the performance seen in non-adversarial settings. A significant portion of this gap can be attributed to the decrease in generalization performance due to adversarial training. In this work, we extend the notion of margin loss to adversarial settings and bound the generalization error for DNNs trained under several well-known gradient-based attack schemes, motivating an effective regularization scheme based on spectral normalization of the DNN’s weight matrices. We also provide a computationally-efficient method for normalizing the spectral norm of convolutional layers with arbitrary stride and padding schemes in deep convolutional networks. We evaluate the power of spectral normalization extensively on combinations of datasets, network architectures, and adversarial training schemes. The code is available at https://github.com/jessemzhang/dl_spectral_normalization. |
Tasks | |
Published | 2018-11-19 |
URL | http://arxiv.org/abs/1811.07457v1 |
http://arxiv.org/pdf/1811.07457v1.pdf | |
PWC | https://paperswithcode.com/paper/generalizable-adversarial-training-via |
Repo | https://github.com/jessemzhang/dl_spectral_normalization |
Framework | tf |
Automatic salt deposits segmentation: A deep learning approach
Title | Automatic salt deposits segmentation: A deep learning approach |
Authors | Mikhail Karchevskiy, Insaf Ashrapov, Leonid Kozinkin |
Abstract | One of the most important applications of seismic reflection is the hydrocarbon exploration which is closely related to salt deposits analysis. This problem is very important even nowadays due to it’s non-linear nature. Taking into account the recent developments in deep learning networks TGS-NOPEC Geophysical Company hosted the Kaggle competition for salt deposits segmentation problem in seismic image data. In this paper, we demonstrate the great performance of several novel deep learning techniques merged into a single neural network which achieved the 27th place (top 1%) in the mentioned competition. Using a U-Net with ResNeXt-50 encoder pre-trained on ImageNet as our base architecture, we implemented Spatial-Channel Squeeze & Excitation, Lovasz loss, CoordConv and Hypercolumn methods. The source code for our solution is made publicly available at https://github.com/K-Mike/Automatic-salt-deposits-segmentation. |
Tasks | |
Published | 2018-11-21 |
URL | http://arxiv.org/abs/1812.01429v1 |
http://arxiv.org/pdf/1812.01429v1.pdf | |
PWC | https://paperswithcode.com/paper/181201429 |
Repo | https://github.com/woans0104/sk_project |
Framework | none |
Lipizzaner: A System That Scales Robust Generative Adversarial Network Training
Title | Lipizzaner: A System That Scales Robust Generative Adversarial Network Training |
Authors | Tom Schmiedlechner, Ignavier Ng Zhi Yong, Abdullah Al-Dujaili, Erik Hemberg, Una-May O’Reilly |
Abstract | GANs are difficult to train due to convergence pathologies such as mode and discriminator collapse. We introduce Lipizzaner, an open source software system that allows machine learning engineers to train GANs in a distributed and robust way. Lipizzaner distributes a competitive coevolutionary algorithm which, by virtue of dual, adapting, generator and discriminator populations, is robust to collapses. The algorithm is well suited to efficient distribution because it uses a spatial grid abstraction. Training is local to each cell and strong intermediate training results are exchanged among overlapping neighborhoods allowing high performing solutions to propagate and improve with more rounds of training. Experiments on common image datasets overcome critical collapses. Communication overhead scales linearly when increasing the number of compute instances and we observe that increasing scale leads to improved model performance. |
Tasks | |
Published | 2018-11-30 |
URL | http://arxiv.org/abs/1811.12843v1 |
http://arxiv.org/pdf/1811.12843v1.pdf | |
PWC | https://paperswithcode.com/paper/lipizzaner-a-system-that-scales-robust |
Repo | https://github.com/ALFA-group/lipizzaner-gan |
Framework | pytorch |
ADCrowdNet: An Attention-injective Deformable Convolutional Network for Crowd Understanding
Title | ADCrowdNet: An Attention-injective Deformable Convolutional Network for Crowd Understanding |
Authors | Ning Liu, Yongchao Long, Changqing Zou, Qun Niu, Li Pan, Hefeng Wu |
Abstract | We propose an attention-injective deformable convolutional network called ADCrowdNet for crowd understanding that can address the accuracy degradation problem of highly congested noisy scenes. ADCrowdNet contains two concatenated networks. An attention-aware network called Attention Map Generator (AMG) first detects crowd regions in images and computes the congestion degree of these regions. Based on detected crowd regions and congestion priors, a multi-scale deformable network called Density Map Estimator (DME) then generates high-quality density maps. With the attention-aware training scheme and multi-scale deformable convolutional scheme, the proposed ADCrowdNet achieves the capability of being more effective to capture the crowd features and more resistant to various noises. We have evaluated our method on four popular crowd counting datasets (ShanghaiTech, UCF_CC_50, WorldEXPO’10, and UCSD) and an extra vehicle counting dataset TRANCOS, and our approach beats existing state-of-the-art approaches on all of these datasets. |
Tasks | Crowd Counting |
Published | 2018-11-29 |
URL | http://arxiv.org/abs/1811.11968v5 |
http://arxiv.org/pdf/1811.11968v5.pdf | |
PWC | https://paperswithcode.com/paper/adcrowdnet-an-attention-injective-deformable |
Repo | https://github.com/BIGKnight/ADCrowd_pytorch_implementation |
Framework | pytorch |
Group Anomaly Detection using Deep Generative Models
Title | Group Anomaly Detection using Deep Generative Models |
Authors | Raghavendra Chalapathy, Edward Toth, Sanjay Chawla |
Abstract | Unlike conventional anomaly detection research that focuses on point anomalies, our goal is to detect anomalous collections of individual data points. In particular, we perform group anomaly detection (GAD) with an emphasis on irregular group distributions (e.g. irregular mixtures of image pixels). GAD is an important task in detecting unusual and anomalous phenomena in real-world applications such as high energy particle physics, social media, and medical imaging. In this paper, we take a generative approach by proposing deep generative models: Adversarial autoencoder (AAE) and variational autoencoder (VAE) for group anomaly detection. Both AAE and VAE detect group anomalies using point-wise input data where group memberships are known a priori. We conduct extensive experiments to evaluate our models on real-world datasets. The empirical results demonstrate that our approach is effective and robust in detecting group anomalies. |
Tasks | Anomaly Detection, Group Anomaly Detection |
Published | 2018-04-13 |
URL | http://arxiv.org/abs/1804.04876v1 |
http://arxiv.org/pdf/1804.04876v1.pdf | |
PWC | https://paperswithcode.com/paper/group-anomaly-detection-using-deep-generative |
Repo | https://github.com/raghavchalapathy/gad |
Framework | none |
Less is more: sampling chemical space with active learning
Title | Less is more: sampling chemical space with active learning |
Authors | Justin S. Smith, Ben Nebgen, Nicholas Lubbers, Olexandr Isayev, Adrian E. Roitberg |
Abstract | The development of accurate and transferable machine learning (ML) potentials for predicting molecular energetics is a challenging task. The process of data generation to train such ML potentials is a task neither well understood nor researched in detail. In this work, we present a fully automated approach for the generation of datasets with the intent of training universal ML potentials. It is based on the concept of active learning (AL) via Query by Committee (QBC), which uses the disagreement between an ensemble of ML potentials to infer the reliability of the ensemble’s prediction. QBC allows the presented AL algorithm to automatically sample regions of chemical space where the ML potential fails to accurately predict the potential energy. AL improves the overall fitness of ANAKIN-ME (ANI) deep learning potentials in rigorous test cases by mitigating human biases in deciding what new training data to use. AL also reduces the training set size to a fraction of the data required when using naive random sampling techniques. To provide validation of our AL approach we develop the COMP6 benchmark (publicly available on GitHub), which contains a diverse set of organic molecules. Through the AL process, it is shown that the AL-based potentials perform as well as the ANI-1 potential on COMP6 with only 10% of the data, and vastly outperforms ANI-1 with 25% the amount of data. Finally, we show that our proposed AL technique develops a universal ANI potential (ANI-1x) that provides accurate energy and force predictions on the entire COMP6 benchmark. This universal ML potential achieves a level of accuracy on par with the best ML potentials for single molecule or materials, while remaining applicable to the general class of organic molecules comprised of the elements CHNO. |
Tasks | Active Learning |
Published | 2018-01-28 |
URL | http://arxiv.org/abs/1801.09319v2 |
http://arxiv.org/pdf/1801.09319v2.pdf | |
PWC | https://paperswithcode.com/paper/less-is-more-sampling-chemical-space-with |
Repo | https://github.com/isayev/ANI1_dataset |
Framework | none |
Solving Atari Games Using Fractals And Entropy
Title | Solving Atari Games Using Fractals And Entropy |
Authors | Sergio Hernandez Cerezo, Guillem Duran Ballester, Spiros Baxevanakis |
Abstract | In this paper, we introduce a novel MCTS based approach that is derived from the laws of the thermodynamics. The algorithm coined Fractal Monte Carlo (FMC), allows us to create an agent that takes intelligent actions in both continuous and discrete environments while providing control over every aspect of the agent behavior. Results show that FMC is several orders of magnitude more efficient than similar techniques, such as MCTS, in the Atari games tested. |
Tasks | Atari Games |
Published | 2018-07-03 |
URL | http://arxiv.org/abs/1807.01081v1 |
http://arxiv.org/pdf/1807.01081v1.pdf | |
PWC | https://paperswithcode.com/paper/solving-atari-games-using-fractals-and |
Repo | https://github.com/FragileTheory/FractalAI |
Framework | none |