Paper Group AWR 59
![Paper Group AWR 59](/2017/images/pwc/paper-arxiv_hu144ec288a26b3e360d673e256787de3e_28623_900x500_fit_q75_box.jpg)
Tracking for Half an Hour. Enhancing Network Embedding with Auxiliary Information: An Explicit Matrix Factorization Perspective. SuperPoint: Self-Supervised Interest Point Detection and Description. A Low-Cost Ethics Shaping Approach for Designing Reinforcement Learning Agents. DropIn: Making Reservoir Computing Neural Networks Robust to Missing In …
Tracking for Half an Hour
Title | Tracking for Half an Hour |
Authors | Ran Tao, Efstratios Gavves, Arnold W. M. Smeulders |
Abstract | Long-term tracking requires extreme stability to the multitude of model updates and robustness to the disappearance and loss of the target as such will inevitably happen. For motivation, we have taken 10 randomly selected OTB-sequences, doubled each by attaching a reversed version and repeated each double sequence 20 times. On most of these repetitive videos, the best current tracker performs worse on each loop. This illustrates the difference between optimization for short-term versus long-term tracking. In a long-term tracker a combined global and local search strategy is beneficial, allowing for recovery from failures and disappearance. Most importantly, the proposed tracker also employs cautious updating, guided by self-quality assessment. The proposed tracker is still among the best on the 20-sec OTB-videos while achieving state-of-the-art on the 100-sec UAV20L benchmark. On 10 new half-an-hour videos with city bicycling, sport games etc, the proposed tracker outperforms others by a large margin where the 2010 TLD tracker comes second. |
Tasks | |
Published | 2017-11-28 |
URL | http://arxiv.org/abs/1711.10217v1 |
http://arxiv.org/pdf/1711.10217v1.pdf | |
PWC | https://paperswithcode.com/paper/tracking-for-half-an-hour |
Repo | https://github.com/QUVA-Lab/Long-term-Siamese-Tracker |
Framework | pytorch |
Enhancing Network Embedding with Auxiliary Information: An Explicit Matrix Factorization Perspective
Title | Enhancing Network Embedding with Auxiliary Information: An Explicit Matrix Factorization Perspective |
Authors | Junliang Guo, Linli Xu, Xunpeng Huang, Enhong Chen |
Abstract | Recent advances in the field of network embedding have shown the low-dimensional network representation is playing a critical role in network analysis. However, most of the existing principles of network embedding do not incorporate auxiliary information such as content and labels of nodes flexibly. In this paper, we take a matrix factorization perspective of network embedding, and incorporate structure, content and label information of the network simultaneously. For structure, we validate that the matrix we construct preserves high-order proximities of the network. Label information can be further integrated into the matrix via the process of random walk sampling to enhance the quality of embedding in an unsupervised manner, i.e., without leveraging downstream classifiers. In addition, we generalize the Skip-Gram Negative Sampling model to integrate the content of the network in a matrix factorization framework. As a consequence, network embedding can be learned in a unified framework integrating network structure and node content as well as label information simultaneously. We demonstrate the efficacy of the proposed model with the tasks of semi-supervised node classification and link prediction on a variety of real-world benchmark network datasets. |
Tasks | Link Prediction, Network Embedding, Node Classification |
Published | 2017-11-11 |
URL | http://arxiv.org/abs/1711.04094v2 |
http://arxiv.org/pdf/1711.04094v2.pdf | |
PWC | https://paperswithcode.com/paper/enhancing-network-embedding-with-auxiliary |
Repo | https://github.com/lemmonation/APNE |
Framework | none |
SuperPoint: Self-Supervised Interest Point Detection and Description
Title | SuperPoint: Self-Supervised Interest Point Detection and Description |
Authors | Daniel DeTone, Tomasz Malisiewicz, Andrew Rabinovich |
Abstract | This paper presents a self-supervised framework for training interest point detectors and descriptors suitable for a large number of multiple-view geometry problems in computer vision. As opposed to patch-based neural networks, our fully-convolutional model operates on full-sized images and jointly computes pixel-level interest point locations and associated descriptors in one forward pass. We introduce Homographic Adaptation, a multi-scale, multi-homography approach for boosting interest point detection repeatability and performing cross-domain adaptation (e.g., synthetic-to-real). Our model, when trained on the MS-COCO generic image dataset using Homographic Adaptation, is able to repeatedly detect a much richer set of interest points than the initial pre-adapted deep model and any other traditional corner detector. The final system gives rise to state-of-the-art homography estimation results on HPatches when compared to LIFT, SIFT and ORB. |
Tasks | Domain Adaptation, Homography Estimation, Interest Point Detection |
Published | 2017-12-20 |
URL | http://arxiv.org/abs/1712.07629v4 |
http://arxiv.org/pdf/1712.07629v4.pdf | |
PWC | https://paperswithcode.com/paper/superpoint-self-supervised-interest-point |
Repo | https://github.com/magicleap/SuperGluePretrainedNetwork |
Framework | pytorch |
A Low-Cost Ethics Shaping Approach for Designing Reinforcement Learning Agents
Title | A Low-Cost Ethics Shaping Approach for Designing Reinforcement Learning Agents |
Authors | Yueh-Hua Wu, Shou-De Lin |
Abstract | This paper proposes a low-cost, easily realizable strategy to equip a reinforcement learning (RL) agent the capability of behaving ethically. Our model allows the designers of RL agents to solely focus on the task to achieve, without having to worry about the implementation of multiple trivial ethical patterns to follow. Based on the assumption that the majority of human behavior, regardless which goals they are achieving, is ethical, our design integrates human policy with the RL policy to achieve the target objective with less chance of violating the ethical code that human beings normally obey. |
Tasks | |
Published | 2017-12-12 |
URL | http://arxiv.org/abs/1712.04172v2 |
http://arxiv.org/pdf/1712.04172v2.pdf | |
PWC | https://paperswithcode.com/paper/a-low-cost-ethics-shaping-approach-for |
Repo | https://github.com/kristery/EthicsShaping |
Framework | none |
DropIn: Making Reservoir Computing Neural Networks Robust to Missing Inputs by Dropout
Title | DropIn: Making Reservoir Computing Neural Networks Robust to Missing Inputs by Dropout |
Authors | Davide Bacciu, Francesco Crecchi, Davide Morelli |
Abstract | The paper presents a novel, principled approach to train recurrent neural networks from the Reservoir Computing family that are robust to missing part of the input features at prediction time. By building on the ensembling properties of Dropout regularization, we propose a methodology, named DropIn, which efficiently trains a neural model as a committee machine of subnetworks, each capable of predicting with a subset of the original input features. We discuss the application of the DropIn methodology in the context of Reservoir Computing models and targeting applications characterized by input sources that are unreliable or prone to be disconnected, such as in pervasive wireless sensor networks and ambient intelligence. We provide an experimental assessment using real-world data from such application domains, showing how the Dropin methodology allows to maintain predictive performances comparable to those of a model without missing features, even when 20%-50% of the inputs are not available. |
Tasks | |
Published | 2017-05-07 |
URL | http://arxiv.org/abs/1705.02643v1 |
http://arxiv.org/pdf/1705.02643v1.pdf | |
PWC | https://paperswithcode.com/paper/dropin-making-reservoir-computing-neural |
Repo | https://github.com/FrancescoCrecchi/DropIn-ESN |
Framework | none |
Can Spatiotemporal 3D CNNs Retrace the History of 2D CNNs and ImageNet?
Title | Can Spatiotemporal 3D CNNs Retrace the History of 2D CNNs and ImageNet? |
Authors | Kensho Hara, Hirokatsu Kataoka, Yutaka Satoh |
Abstract | The purpose of this study is to determine whether current video datasets have sufficient data for training very deep convolutional neural networks (CNNs) with spatio-temporal three-dimensional (3D) kernels. Recently, the performance levels of 3D CNNs in the field of action recognition have improved significantly. However, to date, conventional research has only explored relatively shallow 3D architectures. We examine the architectures of various 3D CNNs from relatively shallow to very deep ones on current video datasets. Based on the results of those experiments, the following conclusions could be obtained: (i) ResNet-18 training resulted in significant overfitting for UCF-101, HMDB-51, and ActivityNet but not for Kinetics. (ii) The Kinetics dataset has sufficient data for training of deep 3D CNNs, and enables training of up to 152 ResNets layers, interestingly similar to 2D ResNets on ImageNet. ResNeXt-101 achieved 78.4% average accuracy on the Kinetics test set. (iii) Kinetics pretrained simple 3D architectures outperforms complex 2D architectures, and the pretrained ResNeXt-101 achieved 94.5% and 70.2% on UCF-101 and HMDB-51, respectively. The use of 2D CNNs trained on ImageNet has produced significant progress in various tasks in image. We believe that using deep 3D CNNs together with Kinetics will retrace the successful history of 2D CNNs and ImageNet, and stimulate advances in computer vision for videos. The codes and pretrained models used in this study are publicly available. https://github.com/kenshohara/3D-ResNets-PyTorch |
Tasks | Action Recognition In Videos |
Published | 2017-11-27 |
URL | http://arxiv.org/abs/1711.09577v2 |
http://arxiv.org/pdf/1711.09577v2.pdf | |
PWC | https://paperswithcode.com/paper/can-spatiotemporal-3d-cnns-retrace-the |
Repo | https://github.com/ptmcgrat/3D-ResNets-PyTorch |
Framework | pytorch |
An Adaptive Genetic Algorithm for Solving N-Queens Problem
Title | An Adaptive Genetic Algorithm for Solving N-Queens Problem |
Authors | Uddalok Sarkar, Sayan Nag |
Abstract | In this paper a Metaheuristic approach for solving the N-Queens Problem is introduced to find the best possible solution in a reasonable amount of time. Genetic Algorithm is used with a novel fitness function as the Metaheuristic. The aim of N-Queens Problem is to place N queens on an N x N chessboard, in a way so that no queen is in conflict with the others. Chromosome representation and genetic operations like Mutation and Crossover are described in detail. Results show that this approach yields promising and satisfactory results in less time compared to that obtained from the previous approaches for several large values of N. |
Tasks | |
Published | 2017-11-01 |
URL | http://arxiv.org/abs/1802.02006v1 |
http://arxiv.org/pdf/1802.02006v1.pdf | |
PWC | https://paperswithcode.com/paper/an-adaptive-genetic-algorithm-for-solving-n |
Repo | https://github.com/depanker/ml-problems |
Framework | none |
Lasso Regularization Paths for NARMAX Models via Coordinate Descent
Title | Lasso Regularization Paths for NARMAX Models via Coordinate Descent |
Authors | Antônio H. Ribeiro, Luis A. Aguirre |
Abstract | We propose a new algorithm for estimating NARMAX models with $L_1$ regularization for models represented as a linear combination of basis functions. Due to the $L_1$-norm penalty the Lasso estimation tends to produce some coefficients that are exactly zero and hence gives interpretable models. The novelty of the contribution is the inclusion of error regressors in the Lasso estimation (which yields a nonlinear regression problem). The proposed algorithm uses cyclical coordinate descent to compute the parameters of the NARMAX models for the entire regularization path. It deals with the error terms by updating the regressor matrix along with the parameter vector. In comparative timings we find that the modification does not reduce the computational efficiency of the original algorithm and can provide the most important regressors in very few inexpensive iterations. The method is illustrated for linear and polynomial models by means of two examples. |
Tasks | |
Published | 2017-10-02 |
URL | http://arxiv.org/abs/1710.00598v2 |
http://arxiv.org/pdf/1710.00598v2.pdf | |
PWC | https://paperswithcode.com/paper/lasso-regularization-paths-for-narmax-models |
Repo | https://github.com/antonior92/NarmaxLasso.jl |
Framework | none |
Parsing Universal Dependencies without training
Title | Parsing Universal Dependencies without training |
Authors | Héctor Martínez Alonso, Željko Agić, Barbara Plank, Anders Søgaard |
Abstract | We propose UDP, the first training-free parser for Universal Dependencies (UD). Our algorithm is based on PageRank and a small set of head attachment rules. It features two-step decoding to guarantee that function words are attached as leaf nodes. The parser requires no training, and it is competitive with a delexicalized transfer system. UDP offers a linguistically sound unsupervised alternative to cross-lingual parsing for UD, which can be used as a baseline for such systems. The parser has very few parameters and is distinctly robust to domain change across languages. |
Tasks | |
Published | 2017-01-11 |
URL | http://arxiv.org/abs/1701.03163v1 |
http://arxiv.org/pdf/1701.03163v1.pdf | |
PWC | https://paperswithcode.com/paper/parsing-universal-dependencies-without |
Repo | https://github.com/hectormartinez/ud_unsup_parser |
Framework | none |
Code Attention: Translating Code to Comments by Exploiting Domain Features
Title | Code Attention: Translating Code to Comments by Exploiting Domain Features |
Authors | Wenhao Zheng, Hong-Yu Zhou, Ming Li, Jianxin Wu |
Abstract | Appropriate comments of code snippets provide insight for code functionality, which are helpful for program comprehension. However, due to the great cost of authoring with the comments, many code projects do not contain adequate comments. Automatic comment generation techniques have been proposed to generate comments from pieces of code in order to alleviate the human efforts in annotating the code. Most existing approaches attempt to exploit certain correlations (usually manually given) between code and generated comments, which could be easily violated if the coding patterns change and hence the performance of comment generation declines. In this paper, we first build C2CGit, a large dataset from open projects in GitHub, which is more than 20$\times$ larger than existing datasets. Then we propose a new attention module called Code Attention to translate code to comments, which is able to utilize the domain features of code snippets, such as symbols and identifiers. We make ablation studies to determine effects of different parts in Code Attention. Experimental results demonstrate that the proposed module has better performance over existing approaches in both BLEU and METEOR. |
Tasks | |
Published | 2017-09-22 |
URL | http://arxiv.org/abs/1709.07642v2 |
http://arxiv.org/pdf/1709.07642v2.pdf | |
PWC | https://paperswithcode.com/paper/code-attention-translating-code-to-comments |
Repo | https://github.com/mf1832146/tree-transformer |
Framework | none |
Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset
Title | Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset |
Authors | Joao Carreira, Andrew Zisserman |
Abstract | The paucity of videos in current action classification datasets (UCF-101 and HMDB-51) has made it difficult to identify good video architectures, as most methods obtain similar performance on existing small-scale benchmarks. This paper re-evaluates state-of-the-art architectures in light of the new Kinetics Human Action Video dataset. Kinetics has two orders of magnitude more data, with 400 human action classes and over 400 clips per class, and is collected from realistic, challenging YouTube videos. We provide an analysis on how current architectures fare on the task of action classification on this dataset and how much performance improves on the smaller benchmark datasets after pre-training on Kinetics. We also introduce a new Two-Stream Inflated 3D ConvNet (I3D) that is based on 2D ConvNet inflation: filters and pooling kernels of very deep image classification ConvNets are expanded into 3D, making it possible to learn seamless spatio-temporal feature extractors from video while leveraging successful ImageNet architecture designs and even their parameters. We show that, after pre-training on Kinetics, I3D models considerably improve upon the state-of-the-art in action classification, reaching 80.9% on HMDB-51 and 98.0% on UCF-101. |
Tasks | Action Classification, Action Recognition In Videos, Skeleton Based Action Recognition |
Published | 2017-05-22 |
URL | http://arxiv.org/abs/1705.07750v3 |
http://arxiv.org/pdf/1705.07750v3.pdf | |
PWC | https://paperswithcode.com/paper/quo-vadis-action-recognition-a-new-model-and |
Repo | https://github.com/ahsaniqbal/Kinetics-FeatureExtractor |
Framework | tf |
Enhancing SDO/HMI images using deep learning
Title | Enhancing SDO/HMI images using deep learning |
Authors | C. J. Diaz Baso, A. Asensio Ramos |
Abstract | The Helioseismic and Magnetic Imager (HMI) provides continuum images and magnetograms with a cadence better than one per minute. It has been continuously observing the Sun 24 hours a day for the past 7 years. The obvious trade-off between full disk observations and spatial resolution makes HMI not enough to analyze the smallest-scale events in the solar atmosphere. Our aim is to develop a new method to enhance HMI data, simultaneously deconvolving and super-resolving images and magnetograms. The resulting images will mimic observations with a diffraction-limited telescope twice the diameter of HMI. Our method, which we call Enhance, is based on two deep fully convolutional neural networks that input patches of HMI observations and output deconvolved and super-resolved data. The neural networks are trained on synthetic data obtained from simulations of the emergence of solar active regions. We have obtained deconvolved and supper-resolved HMI images. To solve this ill-defined problem with infinite solutions we have used a neural network approach to add prior information from the simulations. We test Enhance against Hinode data that has been degraded to a 28 cm diameter telescope showing very good consistency. The code is open source. |
Tasks | |
Published | 2017-06-09 |
URL | http://arxiv.org/abs/1706.02933v2 |
http://arxiv.org/pdf/1706.02933v2.pdf | |
PWC | https://paperswithcode.com/paper/enhancing-sdohmi-images-using-deep-learning |
Repo | https://github.com/cdiazbas/enhance |
Framework | tf |
Robust Synthetic Control
Title | Robust Synthetic Control |
Authors | Muhammad Jehangir Amjad, Devavrat Shah, Dennis Shen |
Abstract | We present a robust generalization of the synthetic control method for comparative case studies. Like the classical method, we present an algorithm to estimate the unobservable counterfactual of a treatment unit. A distinguishing feature of our algorithm is that of de-noising the data matrix via singular value thresholding, which renders our approach robust in multiple facets: it automatically identifies a good subset of donors, overcomes the challenges of missing data, and continues to work well in settings where covariate information may not be provided. To begin, we establish the condition under which the fundamental assumption in synthetic control-like approaches holds, i.e. when the linear relationship between the treatment unit and the donor pool prevails in both the pre- and post-intervention periods. We provide the first finite sample analysis for a broader class of models, the Latent Variable Model, in contrast to Factor Models previously considered in the literature. Further, we show that our de-noising procedure accurately imputes missing entries, producing a consistent estimator of the underlying signal matrix provided $p = \Omega( T^{-1 + \zeta})$ for some $\zeta > 0$; here, $p$ is the fraction of observed data and $T$ is the time interval of interest. Under the same setting, we prove that the mean-squared-error (MSE) in our prediction estimation scales as $O(\sigma^2/p + 1/\sqrt{T})$, where $\sigma^2$ is the noise variance. Using a data aggregation method, we show that the MSE can be made as small as $O(T^{-1/2+\gamma})$ for any $\gamma \in (0, 1/2)$, leading to a consistent estimator. We also introduce a Bayesian framework to quantify the model uncertainty through posterior probabilities. Our experiments, using both real-world and synthetic datasets, demonstrate that our robust generalization yields an improvement over the classical synthetic control method. |
Tasks | |
Published | 2017-11-18 |
URL | http://arxiv.org/abs/1711.06940v1 |
http://arxiv.org/pdf/1711.06940v1.pdf | |
PWC | https://paperswithcode.com/paper/robust-synthetic-control |
Repo | https://github.com/jehangiramjad/tslib |
Framework | none |
Implementing the Deep Q-Network
Title | Implementing the Deep Q-Network |
Authors | Melrose Roderick, James MacGlashan, Stefanie Tellex |
Abstract | The Deep Q-Network proposed by Mnih et al. [2015] has become a benchmark and building point for much deep reinforcement learning research. However, replicating results for complex systems is often challenging since original scientific publications are not always able to describe in detail every important parameter setting and software engineering solution. In this paper, we present results from our work reproducing the results of the DQN paper. We highlight key areas in the implementation that were not covered in great detail in the original paper to make it easier for researchers to replicate these results, including termination conditions and gradient descent algorithms. Finally, we discuss methods for improving the computational performance and provide our own implementation that is designed to work with a range of domains, and not just the original Arcade Learning Environment [Bellemare et al., 2013]. |
Tasks | Atari Games |
Published | 2017-11-20 |
URL | http://arxiv.org/abs/1711.07478v1 |
http://arxiv.org/pdf/1711.07478v1.pdf | |
PWC | https://paperswithcode.com/paper/implementing-the-deep-q-network |
Repo | https://github.com/xgfelicia/Reinforcement-Learning |
Framework | pytorch |
Fraternal Dropout
Title | Fraternal Dropout |
Authors | Konrad Zolna, Devansh Arpit, Dendi Suhubdy, Yoshua Bengio |
Abstract | Recurrent neural networks (RNNs) are important class of architectures among neural networks useful for language modeling and sequential prediction. However, optimizing RNNs is known to be harder compared to feed-forward neural networks. A number of techniques have been proposed in literature to address this problem. In this paper we propose a simple technique called fraternal dropout that takes advantage of dropout to achieve this goal. Specifically, we propose to train two identical copies of an RNN (that share parameters) with different dropout masks while minimizing the difference between their (pre-softmax) predictions. In this way our regularization encourages the representations of RNNs to be invariant to dropout mask, thus being robust. We show that our regularization term is upper bounded by the expectation-linear dropout objective which has been shown to address the gap due to the difference between the train and inference phases of dropout. We evaluate our model and achieve state-of-the-art results in sequence modeling tasks on two benchmark datasets - Penn Treebank and Wikitext-2. We also show that our approach leads to performance improvement by a significant margin in image captioning (Microsoft COCO) and semi-supervised (CIFAR-10) tasks. |
Tasks | Image Captioning, Language Modelling |
Published | 2017-10-31 |
URL | http://arxiv.org/abs/1711.00066v4 |
http://arxiv.org/pdf/1711.00066v4.pdf | |
PWC | https://paperswithcode.com/paper/fraternal-dropout |
Repo | https://github.com/kondiz/fraternal-dropout |
Framework | pytorch |