May 7, 2019

2843 words 14 mins read

Paper Group AWR 11

Paper Group AWR 11

Shallow and Deep Convolutional Networks for Saliency Prediction. Temporal Action Localization in Untrimmed Videos via Multi-stage CNNs. An Efficient Training Algorithm for Kernel Survival Support Vector Machines. Double Thompson Sampling for Dueling Bandits. Temporal Activity Detection in Untrimmed Videos with Recurrent Neural Networks. Causal Lear …

Shallow and Deep Convolutional Networks for Saliency Prediction

Title Shallow and Deep Convolutional Networks for Saliency Prediction
Authors Junting Pan, Kevin McGuinness, Elisa Sayrol, Noel O’Connor, Xavier Giro-i-Nieto
Abstract The prediction of salient areas in images has been traditionally addressed with hand-crafted features based on neuroscience principles. This paper, however, addresses the problem with a completely data-driven approach by training a convolutional neural network (convnet). The learning process is formulated as a minimization of a loss function that measures the Euclidean distance of the predicted saliency map with the provided ground truth. The recent publication of large datasets of saliency prediction has provided enough data to train end-to-end architectures that are both fast and accurate. Two designs are proposed: a shallow convnet trained from scratch, and a another deeper solution whose first three layers are adapted from another network trained for classification. To the authors knowledge, these are the first end-to-end CNNs trained and tested for the purpose of saliency prediction.
Tasks Saliency Prediction
Published 2016-03-02
URL http://arxiv.org/abs/1603.00845v1
PDF http://arxiv.org/pdf/1603.00845v1.pdf
PWC https://paperswithcode.com/paper/shallow-and-deep-convolutional-networks-for
Repo https://github.com/imatge-upc/saliency-2016-cvpr
Framework none

Temporal Action Localization in Untrimmed Videos via Multi-stage CNNs

Title Temporal Action Localization in Untrimmed Videos via Multi-stage CNNs
Authors Zheng Shou, Dongang Wang, Shih-Fu Chang
Abstract We address temporal action localization in untrimmed long videos. This is important because videos in real applications are usually unconstrained and contain multiple action instances plus video content of background scenes or other activities. To address this challenging issue, we exploit the effectiveness of deep networks in temporal action localization via three segment-based 3D ConvNets: (1) a proposal network identifies candidate segments in a long video that may contain actions; (2) a classification network learns one-vs-all action classification model to serve as initialization for the localization network; and (3) a localization network fine-tunes on the learned classification network to localize each action instance. We propose a novel loss function for the localization network to explicitly consider temporal overlap and therefore achieve high temporal localization accuracy. Only the proposal network and the localization network are used during prediction. On two large-scale benchmarks, our approach achieves significantly superior performances compared with other state-of-the-art systems: mAP increases from 1.7% to 7.4% on MEXaction2 and increases from 15.0% to 19.0% on THUMOS 2014, when the overlap threshold for evaluation is set to 0.5.
Tasks Action Classification, Action Localization, Temporal Action Localization, Temporal Localization
Published 2016-01-09
URL http://arxiv.org/abs/1601.02129v2
PDF http://arxiv.org/pdf/1601.02129v2.pdf
PWC https://paperswithcode.com/paper/temporal-action-localization-in-untrimmed
Repo https://github.com/zhengshou/scnn
Framework none

An Efficient Training Algorithm for Kernel Survival Support Vector Machines

Title An Efficient Training Algorithm for Kernel Survival Support Vector Machines
Authors Sebastian Pölsterl, Nassir Navab, Amin Katouzian
Abstract Survival analysis is a fundamental tool in medical research to identify predictors of adverse events and develop systems for clinical decision support. In order to leverage large amounts of patient data, efficient optimisation routines are paramount. We propose an efficient training algorithm for the kernel survival support vector machine (SSVM). We directly optimise the primal objective function and employ truncated Newton optimisation and order statistic trees to significantly lower computational costs compared to previous training algorithms, which require $O(n^4)$ space and $O(p n^6)$ time for datasets with $n$ samples and $p$ features. Our results demonstrate that our proposed optimisation scheme allows analysing data of a much larger scale with no loss in prediction performance. Experiments on synthetic and 5 real-world datasets show that our technique outperforms existing kernel SSVM formulations if the amount of right censoring is high ($\geq85%$), and performs comparably otherwise.
Tasks Survival Analysis
Published 2016-11-21
URL http://arxiv.org/abs/1611.07054v1
PDF http://arxiv.org/pdf/1611.07054v1.pdf
PWC https://paperswithcode.com/paper/an-efficient-training-algorithm-for-kernel
Repo https://github.com/sebp/scikit-survival
Framework none

Double Thompson Sampling for Dueling Bandits

Title Double Thompson Sampling for Dueling Bandits
Authors Huasen Wu, Xin Liu
Abstract In this paper, we propose a Double Thompson Sampling (D-TS) algorithm for dueling bandit problems. As indicated by its name, D-TS selects both the first and the second candidates according to Thompson Sampling. Specifically, D-TS maintains a posterior distribution for the preference matrix, and chooses the pair of arms for comparison by sampling twice from the posterior distribution. This simple algorithm applies to general Copeland dueling bandits, including Condorcet dueling bandits as its special case. For general Copeland dueling bandits, we show that D-TS achieves $O(K^2 \log T)$ regret. For Condorcet dueling bandits, we further simplify the D-TS algorithm and show that the simplified D-TS algorithm achieves $O(K \log T + K^2 \log \log T)$ regret. Simulation results based on both synthetic and real-world data demonstrate the efficiency of the proposed D-TS algorithm.
Tasks
Published 2016-04-25
URL http://arxiv.org/abs/1604.07101v2
PDF http://arxiv.org/pdf/1604.07101v2.pdf
PWC https://paperswithcode.com/paper/double-thompson-sampling-for-dueling-bandits
Repo https://github.com/HuasenWu/DuelingBandits
Framework none

Temporal Activity Detection in Untrimmed Videos with Recurrent Neural Networks

Title Temporal Activity Detection in Untrimmed Videos with Recurrent Neural Networks
Authors Alberto Montes, Amaia Salvador, Santiago Pascual, Xavier Giro-i-Nieto
Abstract This thesis explore different approaches using Convolutional and Recurrent Neural Networks to classify and temporally localize activities on videos, furthermore an implementation to achieve it has been proposed. As the first step, features have been extracted from video frames using an state of the art 3D Convolutional Neural Network. This features are fed in a recurrent neural network that solves the activity classification and temporally location tasks in a simple and flexible way. Different architectures and configurations have been tested in order to achieve the best performance and learning of the video dataset provided. In addition it has been studied different kind of post processing over the trained network’s output to achieve a better results on the temporally localization of activities on the videos. The results provided by the neural network developed in this thesis have been submitted to the ActivityNet Challenge 2016 of the CVPR, achieving competitive results using a simple and flexible architecture.
Tasks Action Detection, Activity Detection
Published 2016-08-29
URL http://arxiv.org/abs/1608.08128v3
PDF http://arxiv.org/pdf/1608.08128v3.pdf
PWC https://paperswithcode.com/paper/temporal-activity-detection-in-untrimmed
Repo https://github.com/vohoaiviet/activitynet-2016-cvprw
Framework tf

Causal Learning via Manifold Regularization

Title Causal Learning via Manifold Regularization
Authors Steven M. Hill, Chris. J. Oates, Duncan A. Blythe, Sach Mukherjee
Abstract This paper frames causal structure estimation as a machine learning task. The idea is to treat indicators of causal relationships between variables as `labels’ and to exploit available data on the variables of interest to provide features for the labelling task. Background scientific knowledge or any available interventional data provide labels on some causal relationships and the remainder are treated as unlabelled. To illustrate the key ideas, we develop a distance-based approach (based on bivariate histograms) within a manifold regularization framework. We present empirical results on three different biological data sets (including examples where causal effects can be verified by experimental intervention), that together demonstrate the efficacy and general nature of the approach as well as its simplicity from a user’s point of view. |
Tasks Causal Discovery
Published 2016-12-16
URL https://arxiv.org/abs/1612.05678v4
PDF https://arxiv.org/pdf/1612.05678v4.pdf
PWC https://paperswithcode.com/paper/causal-discovery-as-semi-supervised-learning
Repo https://github.com/Steven-M-Hill/MRCL
Framework none

Equilibrium Propagation: Bridging the Gap Between Energy-Based Models and Backpropagation

Title Equilibrium Propagation: Bridging the Gap Between Energy-Based Models and Backpropagation
Authors Benjamin Scellier, Yoshua Bengio
Abstract We introduce Equilibrium Propagation, a learning framework for energy-based models. It involves only one kind of neural computation, performed in both the first phase (when the prediction is made) and the second phase of training (after the target or prediction error is revealed). Although this algorithm computes the gradient of an objective function just like Backpropagation, it does not need a special computation or circuit for the second phase, where errors are implicitly propagated. Equilibrium Propagation shares similarities with Contrastive Hebbian Learning and Contrastive Divergence while solving the theoretical issues of both algorithms: our algorithm computes the gradient of a well defined objective function. Because the objective function is defined in terms of local perturbations, the second phase of Equilibrium Propagation corresponds to only nudging the prediction (fixed point, or stationary distribution) towards a configuration that reduces prediction error. In the case of a recurrent multi-layer supervised network, the output units are slightly nudged towards their target in the second phase, and the perturbation introduced at the output layer propagates backward in the hidden layers. We show that the signal ‘back-propagated’ during this second phase corresponds to the propagation of error derivatives and encodes the gradient of the objective function, when the synaptic update corresponds to a standard form of spike-timing dependent plasticity. This work makes it more plausible that a mechanism similar to Backpropagation could be implemented by brains, since leaky integrator neural computation performs both inference and error back-propagation in our model. The only local difference between the two phases is whether synaptic changes are allowed or not.
Tasks
Published 2016-02-16
URL http://arxiv.org/abs/1602.05179v5
PDF http://arxiv.org/pdf/1602.05179v5.pdf
PWC https://paperswithcode.com/paper/equilibrium-propagation-bridging-the-gap
Repo https://github.com/bscellier/Towards-a-Biologically-Plausible-Backprop
Framework pytorch

Ternary Weight Networks

Title Ternary Weight Networks
Authors Fengfu Li, Bo Zhang, Bin Liu
Abstract We introduce ternary weight networks (TWNs) - neural networks with weights constrained to +1, 0 and -1. The Euclidian distance between full (float or double) precision weights and the ternary weights along with a scaling factor is minimized. Besides, a threshold-based ternary function is optimized to get an approximated solution which can be fast and easily computed. TWNs have stronger expressive abilities than the recently proposed binary precision counterparts and are thus more effective than the latter. Meanwhile, TWNs achieve up to 16$\times$ or 32$\times$ model compression rate and need fewer multiplications compared with the full precision counterparts. Benchmarks on MNIST, CIFAR-10, and large scale ImageNet datasets show that the performance of TWNs is only slightly worse than the full precision counterparts but outperforms the analogous binary precision counterparts a lot.
Tasks Model Compression
Published 2016-05-16
URL http://arxiv.org/abs/1605.04711v2
PDF http://arxiv.org/pdf/1605.04711v2.pdf
PWC https://paperswithcode.com/paper/ternary-weight-networks
Repo https://github.com/da-steve101/binary_connect_cifar
Framework none

Minimizing Quadratic Functions in Constant Time

Title Minimizing Quadratic Functions in Constant Time
Authors Kohei Hayashi, Yuichi Yoshida
Abstract A sampling-based optimization method for quadratic functions is proposed. Our method approximately solves the following $n$-dimensional quadratic minimization problem in constant time, which is independent of $n$: $z^*=\min_{\mathbf{v} \in \mathbb{R}^n}\langle\mathbf{v}, A \mathbf{v}\rangle + n\langle\mathbf{v}, \mathrm{diag}(\mathbf{d})\mathbf{v}\rangle + n\langle\mathbf{b}, \mathbf{v}\rangle$, where $A \in \mathbb{R}^{n \times n}$ is a matrix and $\mathbf{d},\mathbf{b} \in \mathbb{R}^n$ are vectors. Our theoretical analysis specifies the number of samples $k(\delta, \epsilon)$ such that the approximated solution $z$ satisfies $z - z^* = O(\epsilon n^2)$ with probability $1-\delta$. The empirical performance (accuracy and runtime) is positively confirmed by numerical experiments.
Tasks
Published 2016-08-25
URL http://arxiv.org/abs/1608.07179v1
PDF http://arxiv.org/pdf/1608.07179v1.pdf
PWC https://paperswithcode.com/paper/minimizing-quadratic-functions-in-constant
Repo https://github.com/hayasick/CTOQ
Framework none

Adversarial Training Methods for Semi-Supervised Text Classification

Title Adversarial Training Methods for Semi-Supervised Text Classification
Authors Takeru Miyato, Andrew M. Dai, Ian Goodfellow
Abstract Adversarial training provides a means of regularizing supervised learning algorithms while virtual adversarial training is able to extend supervised learning algorithms to the semi-supervised setting. However, both methods require making small perturbations to numerous entries of the input vector, which is inappropriate for sparse high-dimensional inputs such as one-hot word representations. We extend adversarial and virtual adversarial training to the text domain by applying perturbations to the word embeddings in a recurrent neural network rather than to the original input itself. The proposed method achieves state of the art results on multiple benchmark semi-supervised and purely supervised tasks. We provide visualizations and analysis showing that the learned word embeddings have improved in quality and that while training, the model is less prone to overfitting.
Tasks Sentiment Analysis, Text Classification, Word Embeddings
Published 2016-05-25
URL http://arxiv.org/abs/1605.07725v3
PDF http://arxiv.org/pdf/1605.07725v3.pdf
PWC https://paperswithcode.com/paper/adversarial-training-methods-for-semi
Repo https://github.com/TobiasLee/Text-Classification
Framework tf

Relation Schema Induction using Tensor Factorization with Side Information

Title Relation Schema Induction using Tensor Factorization with Side Information
Authors Madhav Nimishakavi, Uday Singh Saini, Partha Talukdar
Abstract Given a set of documents from a specific domain (e.g., medical research journals), how do we automatically build a Knowledge Graph (KG) for that domain? Automatic identification of relations and their schemas, i.e., type signature of arguments of relations (e.g., undergo(Patient, Surgery)), is an important first step towards this goal. We refer to this problem as Relation Schema Induction (RSI). In this paper, we propose Schema Induction using Coupled Tensor Factorization (SICTF), a novel tensor factorization method for relation schema induction. SICTF factorizes Open Information Extraction (OpenIE) triples extracted from a domain corpus along with additional side information in a principled way to induce relation schemas. To the best of our knowledge, this is the first application of tensor factorization for the RSI problem. Through extensive experiments on multiple real-world datasets, we find that SICTF is not only more accurate than state-of-the-art baselines, but also significantly faster (about 14x faster).
Tasks Open Information Extraction
Published 2016-05-12
URL http://arxiv.org/abs/1605.04227v3
PDF http://arxiv.org/pdf/1605.04227v3.pdf
PWC https://paperswithcode.com/paper/relation-schema-induction-using-tensor
Repo https://github.com/malllabiisc/sictf
Framework none

Quantum Algorithms for Compositional Natural Language Processing

Title Quantum Algorithms for Compositional Natural Language Processing
Authors William Zeng, Bob Coecke
Abstract We propose a new application of quantum computing to the field of natural language processing. Ongoing work in this field attempts to incorporate grammatical structure into algorithms that compute meaning. In (Coecke, Sadrzadeh and Clark, 2010), the authors introduce such a model (the CSC model) based on tensor product composition. While this algorithm has many advantages, its implementation is hampered by the large classical computational resources that it requires. In this work we show how computational shortcomings of the CSC approach could be resolved using quantum computation (possibly in addition to existing techniques for dimension reduction). We address the value of quantum RAM (Giovannetti,2008) for this model and extend an algorithm from Wiebe, Braun and Lloyd (2012) into a quantum algorithm to categorize sentences in CSC. Our new algorithm demonstrates a quadratic speedup over classical methods under certain conditions.
Tasks Dimensionality Reduction
Published 2016-08-04
URL http://arxiv.org/abs/1608.01406v1
PDF http://arxiv.org/pdf/1608.01406v1.pdf
PWC https://paperswithcode.com/paper/quantum-algorithms-for-compositional-natural
Repo https://github.com/ICHEC/QNLP
Framework none

Multiple Object Tracking with Kernelized Correlation Filters in Urban Mixed Traffic

Title Multiple Object Tracking with Kernelized Correlation Filters in Urban Mixed Traffic
Authors Yuebin Yang, Guillaume-Alexandre Bilodeau
Abstract Recently, the Kernelized Correlation Filters tracker (KCF) achieved competitive performance and robustness in visual object tracking. On the other hand, visual trackers are not typically used in multiple object tracking. In this paper, we investigate how a robust visual tracker like KCF can improve multiple object tracking. Since KCF is a fast tracker, many can be used in parallel and still result in fast tracking. We build a multiple object tracking system based on KCF and background subtraction. Background subtraction is applied to extract moving objects and get their scale and size in combination with KCF outputs, while KCF is used for data association and to handle fragmentation and occlusion problems. As a result, KCF and background subtraction help each other to take tracking decision at every frame. Sometimes KCF outputs are the most trustworthy (e.g. during occlusion), while in some other case, it is the background subtraction outputs. To validate the effectiveness of our system, the algorithm is demonstrated on four urban video recordings from a standard dataset. Results show that our method is competitive with state-of-the-art trackers even if we use a much simpler data association step.
Tasks Multiple Object Tracking, Object Tracking, Visual Object Tracking
Published 2016-11-08
URL http://arxiv.org/abs/1611.02364v2
PDF http://arxiv.org/pdf/1611.02364v2.pdf
PWC https://paperswithcode.com/paper/multiple-object-tracking-with-kernelized
Repo https://github.com/iyybpatrick/MKCF
Framework none

Fitting a Simplicial Complex using a Variation of k-means

Title Fitting a Simplicial Complex using a Variation of k-means
Authors Piotr Beben
Abstract We give a simple and effective two stage algorithm for approximating a point cloud $\mathcal{S}\subset\mathbb{R}^m$ by a simplicial complex $K$. The first stage is an iterative fitting procedure that generalizes k-means clustering, while the second stage involves deleting redundant simplices. A form of dimension reduction of $\mathcal{S}$ is obtained as a consequence.
Tasks Dimensionality Reduction
Published 2016-07-13
URL http://arxiv.org/abs/1607.03849v2
PDF http://arxiv.org/pdf/1607.03849v2.pdf
PWC https://paperswithcode.com/paper/fitting-a-simplicial-complex-using-a
Repo https://github.com/pbebenSoton/smeans
Framework none

Character-Level Neural Translation for Multilingual Media Monitoring in the SUMMA Project

Title Character-Level Neural Translation for Multilingual Media Monitoring in the SUMMA Project
Authors Guntis Barzdins, Steve Renals, Didzis Gosko
Abstract The paper steps outside the comfort-zone of the traditional NLP tasks like automatic speech recognition (ASR) and machine translation (MT) to addresses two novel problems arising in the automated multilingual news monitoring: segmentation of the TV and radio program ASR transcripts into individual stories, and clustering of the individual stories coming from various sources and languages into storylines. Storyline clustering of stories covering the same events is an essential task for inquisitorial media monitoring. We address these two problems jointly by engaging the low-dimensional semantic representation capabilities of the sequence to sequence neural translation models. To enable joint multi-task learning for multilingual neural translation of morphologically rich languages we replace the attention mechanism with the sliding-window mechanism and operate the sequence to sequence neural translation model on the character-level rather than on the word-level. The story segmentation and storyline clustering problem is tackled by examining the low-dimensional vectors produced as a side-product of the neural translation process. The results of this paper describe a novel approach to the automatic story segmentation and storyline clustering problem.
Tasks Machine Translation, Multi-Task Learning, Speech Recognition
Published 2016-04-05
URL http://arxiv.org/abs/1604.01221v1
PDF http://arxiv.org/pdf/1604.01221v1.pdf
PWC https://paperswithcode.com/paper/character-level-neural-translation-for
Repo https://github.com/didzis/tensorflowAMR
Framework tf
comments powered by Disqus