May 7, 2019

2746 words 13 mins read

Paper Group ANR 89

Paper Group ANR 89

Byte-based Language Identification with Deep Convolutional Networks. Hierarchical Conflict Propagation: Sequence Learning in a Recurrent Deep Neural Network. Fine-to-coarse Knowledge Transfer For Low-Res Image Classification. Photographic dataset: random peppercorns. Solving MaxSAT by Successive Calls to a SAT Solver. Variational Autoencoder for De …

Byte-based Language Identification with Deep Convolutional Networks

Title Byte-based Language Identification with Deep Convolutional Networks
Authors Johannes Bjerva
Abstract We report on our system for the shared task on discriminating between similar languages (DSL 2016). The system uses only byte representations in a deep residual network (ResNet). The system, named ResIdent, is trained only on the data released with the task (closed training). We obtain 84.88% accuracy on subtask A, 68.80% accuracy on subtask B1, and 69.80% accuracy on subtask B2. A large difference in accuracy on development data can be observed with relatively minor changes in our network’s architecture and hyperparameters. We therefore expect fine-tuning of these parameters to yield higher accuracies.
Tasks Language Identification
Published 2016-09-28
URL http://arxiv.org/abs/1609.09004v2
PDF http://arxiv.org/pdf/1609.09004v2.pdf
PWC https://paperswithcode.com/paper/byte-based-language-identification-with-deep
Repo
Framework

Hierarchical Conflict Propagation: Sequence Learning in a Recurrent Deep Neural Network

Title Hierarchical Conflict Propagation: Sequence Learning in a Recurrent Deep Neural Network
Authors Andrew J. R. Simpson
Abstract Recurrent neural networks (RNN) are capable of learning to encode and exploit activation history over an arbitrary timescale. However, in practice, state of the art gradient descent based training methods are known to suffer from difficulties in learning long term dependencies. Here, we describe a novel training method that involves concurrent parallel cloned networks, each sharing the same weights, each trained at different stimulus phase and each maintaining independent activation histories. Training proceeds by recursively performing batch-updates over the parallel clones as activation history is progressively increased. This allows conflicts to propagate hierarchically from short-term contexts towards longer-term contexts until they are resolved. We illustrate the parallel clones method and hierarchical conflict propagation with a character-level deep RNN tasked with memorizing a paragraph of Moby Dick (by Herman Melville).
Tasks
Published 2016-02-25
URL http://arxiv.org/abs/1602.08118v1
PDF http://arxiv.org/pdf/1602.08118v1.pdf
PWC https://paperswithcode.com/paper/hierarchical-conflict-propagation-sequence
Repo
Framework

Fine-to-coarse Knowledge Transfer For Low-Res Image Classification

Title Fine-to-coarse Knowledge Transfer For Low-Res Image Classification
Authors Xingchao Peng, Judy Hoffman, Stella X. Yu, Kate Saenko
Abstract We address the difficult problem of distinguishing fine-grained object categories in low resolution images. Wepropose a simple an effective deep learning approach that transfers fine-grained knowledge gained from high resolution training data to the coarse low-resolution test scenario. Such fine-to-coarse knowledge transfer has many real world applications, such as identifying objects in surveillance photos or satellite images where the image resolution at the test time is very low but plenty of high resolution photos of similar objects are available. Our extensive experiments on two standard benchmark datasets containing fine-grained car models and bird species demonstrate that our approach can effectively transfer fine-detail knowledge to coarse-detail imagery.
Tasks Image Classification, Transfer Learning
Published 2016-05-21
URL http://arxiv.org/abs/1605.06695v1
PDF http://arxiv.org/pdf/1605.06695v1.pdf
PWC https://paperswithcode.com/paper/fine-to-coarse-knowledge-transfer-for-low-res
Repo
Framework

Photographic dataset: random peppercorns

Title Photographic dataset: random peppercorns
Authors Teemu Helenius, Samuli Siltanen
Abstract This is a photographic dataset collected for testing image processing algorithms. The idea is to have sets of different but statistically similar images. In this work the images show randomly distributed peppercorns. The dataset is made available at www.fips.fi/photographic_dataset.php .
Tasks
Published 2016-03-03
URL http://arxiv.org/abs/1603.01046v1
PDF http://arxiv.org/pdf/1603.01046v1.pdf
PWC https://paperswithcode.com/paper/photographic-dataset-random-peppercorns
Repo
Framework

Solving MaxSAT by Successive Calls to a SAT Solver

Title Solving MaxSAT by Successive Calls to a SAT Solver
Authors Mohamed El Halaby
Abstract The Maximum Satisfiability (MaxSAT) problem is the problem of finding a truth assignment that maximizes the number of satisfied clauses of a given Boolean formula in Conjunctive Normal Form (CNF). Many exact solvers for MaxSAT have been developed during recent years, and many of them were presented in the well-known SAT conference. Algorithms for MaxSAT generally fall into two categories: (1) branch and bound algorithms and (2) algorithms that use successive calls to a SAT solver (SAT- based), which this paper in on. In practical problems, SAT-based algorithms have been shown to be more efficient. This paper provides an experimental investigation to compare the performance of recent SAT-based and branch and bound algorithms on the benchmarks of the MaxSAT Evaluations.
Tasks
Published 2016-03-11
URL http://arxiv.org/abs/1603.03814v1
PDF http://arxiv.org/pdf/1603.03814v1.pdf
PWC https://paperswithcode.com/paper/solving-maxsat-by-successive-calls-to-a-sat
Repo
Framework

Variational Autoencoder for Deep Learning of Images, Labels and Captions

Title Variational Autoencoder for Deep Learning of Images, Labels and Captions
Authors Yunchen Pu, Zhe Gan, Ricardo Henao, Xin Yuan, Chunyuan Li, Andrew Stevens, Lawrence Carin
Abstract A novel variational autoencoder is developed to model images, as well as associated labels or captions. The Deep Generative Deconvolutional Network (DGDN) is used as a decoder of the latent image features, and a deep Convolutional Neural Network (CNN) is used as an image encoder; the CNN is used to approximate a distribution for the latent DGDN features/code. The latent code is also linked to generative models for labels (Bayesian support vector machine) or captions (recurrent neural network). When predicting a label/caption for a new image at test, averaging is performed across the distribution of latent codes; this is computationally efficient as a consequence of the learned CNN-based encoder. Since the framework is capable of modeling the image in the presence/absence of associated labels/captions, a new semi-supervised setting is manifested for CNN learning with images; the framework even allows unsupervised CNN learning, based on images alone.
Tasks
Published 2016-09-28
URL http://arxiv.org/abs/1609.08976v1
PDF http://arxiv.org/pdf/1609.08976v1.pdf
PWC https://paperswithcode.com/paper/variational-autoencoder-for-deep-learning-of
Repo
Framework

CUNI System for WMT16 Automatic Post-Editing and Multimodal Translation Tasks

Title CUNI System for WMT16 Automatic Post-Editing and Multimodal Translation Tasks
Authors Jindřich Libovický, Jindřich Helcl, Marek Tlustý, Pavel Pecina, Ondřej Bojar
Abstract Neural sequence to sequence learning recently became a very promising paradigm in machine translation, achieving competitive results with statistical phrase-based systems. In this system description paper, we attempt to utilize several recently published methods used for neural sequential learning in order to build systems for WMT 2016 shared tasks of Automatic Post-Editing and Multimodal Machine Translation.
Tasks Automatic Post-Editing, Machine Translation, Multimodal Machine Translation
Published 2016-06-23
URL http://arxiv.org/abs/1606.07481v1
PDF http://arxiv.org/pdf/1606.07481v1.pdf
PWC https://paperswithcode.com/paper/cuni-system-for-wmt16-automatic-post-editing
Repo
Framework

Psychologically based Virtual-Suspect for Interrogative Interview Training

Title Psychologically based Virtual-Suspect for Interrogative Interview Training
Authors Moshe Bitan, Galit Nahari, Zvi Nisin, Ariel Roth, Sarit Kraus
Abstract In this paper, we present a Virtual-Suspect system which can be used to train inexperienced law enforcement personnel in interrogation strategies. The system supports different scenario configurations based on historical data. The responses presented by the Virtual-Suspect are selected based on the psychological state of the suspect, which can be configured as well. Furthermore, each interrogator’s statement affects the Virtual-Suspect’s current psychological state, which may lead the interrogation in different directions. In addition, the model takes into account the context in which the statements are made. Experiments with 24 subjects demonstrate that the Virtual-Suspect’s behavior is similar to that of a human who plays the role of the suspect.
Tasks
Published 2016-05-31
URL http://arxiv.org/abs/1605.09505v1
PDF http://arxiv.org/pdf/1605.09505v1.pdf
PWC https://paperswithcode.com/paper/psychologically-based-virtual-suspect-for
Repo
Framework

Collaborative Learning of Stochastic Bandits over a Social Network

Title Collaborative Learning of Stochastic Bandits over a Social Network
Authors Ravi Kumar Kolla, Krishna Jagannathan, Aditya Gopalan
Abstract We consider a collaborative online learning paradigm, wherein a group of agents connected through a social network are engaged in playing a stochastic multi-armed bandit game. Each time an agent takes an action, the corresponding reward is instantaneously observed by the agent, as well as its neighbours in the social network. We perform a regret analysis of various policies in this collaborative learning setting. A key finding of this paper is that natural extensions of widely-studied single agent learning policies to the network setting need not perform well in terms of regret. In particular, we identify a class of non-altruistic and individually consistent policies, and argue by deriving regret lower bounds that they are liable to suffer a large regret in the networked setting. We also show that the learning performance can be substantially improved if the agents exploit the structure of the network, and develop a simple learning algorithm based on dominating sets of the network. Specifically, we first consider a star network, which is a common motif in hierarchical social networks, and show analytically that the hub agent can be used as an information sink to expedite learning and improve the overall regret. We also derive networkwide regret bounds for the algorithm applied to general networks. We conduct numerical experiments on a variety of networks to corroborate our analytical results.
Tasks
Published 2016-02-29
URL http://arxiv.org/abs/1602.08886v2
PDF http://arxiv.org/pdf/1602.08886v2.pdf
PWC https://paperswithcode.com/paper/collaborative-learning-of-stochastic-bandits
Repo
Framework

Multi-Camera Action Dataset for Cross-Camera Action Recognition Benchmarking

Title Multi-Camera Action Dataset for Cross-Camera Action Recognition Benchmarking
Authors Wenhui Li, Yongkang Wong, An-An Liu, Yang Li, Yu-Ting Su, Mohan Kankanhalli
Abstract Action recognition has received increasing attention from the computer vision and machine learning communities in the last decade. To enable the study of this problem, there exist a vast number of action datasets, which are recorded under controlled laboratory settings, real-world surveillance environments, or crawled from the Internet. Apart from the “in-the-wild” datasets, the training and test split of conventional datasets often possess similar environments conditions, which leads to close to perfect performance on constrained datasets. In this paper, we introduce a new dataset, namely Multi-Camera Action Dataset (MCAD), which is designed to evaluate the open view classification problem under the surveillance environment. In total, MCAD contains 14,298 action samples from 18 action categories, which are performed by 20 subjects and independently recorded with 5 cameras. Inspired by the well received evaluation approach on the LFW dataset, we designed a standard evaluation protocol and benchmarked MCAD under several scenarios. The benchmark shows that while an average of 85% accuracy is achieved under the closed-view scenario, the performance suffers from a significant drop under the cross-view scenario. In the worst case scenario, the performance of 10-fold cross validation drops from 87.0% to 47.4%.
Tasks Temporal Action Localization
Published 2016-07-21
URL http://arxiv.org/abs/1607.06408v3
PDF http://arxiv.org/pdf/1607.06408v3.pdf
PWC https://paperswithcode.com/paper/multi-camera-action-dataset-for-cross-camera
Repo
Framework

Leveraging Over Priors for Boosting Control of Prosthetic Hands

Title Leveraging Over Priors for Boosting Control of Prosthetic Hands
Authors Valentina Gregori
Abstract The Electromyography (EMG) signal is the electrical activity produced by cells of skeletal muscles in order to provide a movement. The non-invasive prosthetic hand works with several electrodes, placed on the stump of an amputee, that record this signal. In order to favour the control of prosthesis, the EMG signal is analyzed with algorithms based on machine learning theory to decide the movement that the subject is going to do. In order to obtain a significant control of the prosthesis and avoid mismatch between desired and performed movements, a long training period is needed when we use the traditional algorithm of machine learning (i.e. Support Vector Machines). An actual challenge in this field concerns the reduction of the time necessary for an amputee to learn how to use the prosthesis. Recently, several algorithms that exploit a form of prior knowledge have been proposed. In general, we refer to prior knowledge as a past experience available in the form of models. In our case an amputee, that attempts to perform some movements with the prosthesis, could use experience from different subjects that are already able to perform those movements. The aim of this work is to verify, with a computational investigation, if for an amputee this kind of previous experience is useful in order to reduce the training time and boost the prosthetic control. Furthermore, we want to understand if and how the final results change when the previous knowledge of intact or amputated subjects is used for a new amputee. Our experiments indicate that: (1) the use of experience, from other subjects already trained to perform a task, makes us able to reduce the training time of about an order of magnitude; (2) it seems that an amputee that tries to learn to use the prosthesis doesn’t reach different results when he/she exploits previous experience of amputees or intact.
Tasks Electromyography (EMG)
Published 2016-05-24
URL http://arxiv.org/abs/1605.07498v1
PDF http://arxiv.org/pdf/1605.07498v1.pdf
PWC https://paperswithcode.com/paper/leveraging-over-priors-for-boosting-control
Repo
Framework

Revisiting Winner Take All (WTA) Hashing for Sparse Datasets

Title Revisiting Winner Take All (WTA) Hashing for Sparse Datasets
Authors Beidi Chen, Anshumali Shrivastava
Abstract WTA (Winner Take All) hashing has been successfully applied in many large scale vision applications. This hashing scheme was tailored to take advantage of the comparative reasoning (or order based information), which showed significant accuracy improvements. In this paper, we identify a subtle issue with WTA, which grows with the sparsity of the datasets. This issue limits the discriminative power of WTA. We then propose a solution for this problem based on the idea of Densification which provably fixes the issue. Our experiments show that Densified WTA Hashing outperforms Vanilla WTA both in image classification and retrieval tasks consistently and significantly.
Tasks Image Classification
Published 2016-12-06
URL http://arxiv.org/abs/1612.01834v2
PDF http://arxiv.org/pdf/1612.01834v2.pdf
PWC https://paperswithcode.com/paper/revisiting-winner-take-all-wta-hashing-for
Repo
Framework

Alternating Back-Propagation for Generator Network

Title Alternating Back-Propagation for Generator Network
Authors Tian Han, Yang Lu, Song-Chun Zhu, Ying Nian Wu
Abstract This paper proposes an alternating back-propagation algorithm for learning the generator network model. The model is a non-linear generalization of factor analysis. In this model, the mapping from the continuous latent factors to the observed signal is parametrized by a convolutional neural network. The alternating back-propagation algorithm iterates the following two steps: (1) Inferential back-propagation, which infers the latent factors by Langevin dynamics or gradient descent. (2) Learning back-propagation, which updates the parameters given the inferred latent factors by gradient descent. The gradient computations in both steps are powered by back-propagation, and they share most of their code in common. We show that the alternating back-propagation algorithm can learn realistic generator models of natural images, video sequences, and sounds. Moreover, it can also be used to learn from incomplete or indirect training data.
Tasks
Published 2016-06-28
URL http://arxiv.org/abs/1606.08571v4
PDF http://arxiv.org/pdf/1606.08571v4.pdf
PWC https://paperswithcode.com/paper/alternating-back-propagation-for-generator
Repo
Framework

Estimation of Fiber Orientations Using Neighborhood Information

Title Estimation of Fiber Orientations Using Neighborhood Information
Authors Chuyang Ye, Jiachen Zhuo, Rao P. Gullapalli, Jerry L. Prince
Abstract Data from diffusion magnetic resonance imaging (dMRI) can be used to reconstruct fiber tracts, for example, in muscle and white matter. Estimation of fiber orientations (FOs) is a crucial step in the reconstruction process and these estimates can be corrupted by noise. In this paper, a new method called Fiber Orientation Reconstruction using Neighborhood Information (FORNI) is described and shown to reduce the effects of noise and improve FO estimation performance by incorporating spatial consistency. FORNI uses a fixed tensor basis to model the diffusion weighted signals, which has the advantage of providing an explicit relationship between the basis vectors and the FOs. FO spatial coherence is encouraged using weighted l1-norm regularization terms, which contain the interaction of directional information between neighbor voxels. Data fidelity is encouraged using a squared error between the observed and reconstructed diffusion weighted signals. After appropriate weighting of these competing objectives, the resulting objective function is minimized using a block coordinate descent algorithm, and a straightforward parallelization strategy is used to speed up processing. Experiments were performed on a digital crossing phantom, ex vivo tongue dMRI data, and in vivo brain dMRI data for both qualitative and quantitative evaluation. The results demonstrate that FORNI improves the quality of FO estimation over other state of the art algorithms.
Tasks
Published 2016-01-16
URL http://arxiv.org/abs/1601.04115v2
PDF http://arxiv.org/pdf/1601.04115v2.pdf
PWC https://paperswithcode.com/paper/estimation-of-fiber-orientations-using
Repo
Framework

Double-detector for Sparse Signal Detection from One Bit Compressed Sensing Measurements

Title Double-detector for Sparse Signal Detection from One Bit Compressed Sensing Measurements
Authors Hadi Zayyani, Farzan Haddadi, Mehdi Korki
Abstract This letter presents the sparse vector signal detection from one bit compressed sensing measurements, in contrast to the previous works which deal with scalar signal detection. In this letter, available results are extended to the vector case and the GLRT detector and the optimal quantizer design are obtained. Also, a double-detector scheme is introduced in which a sensor level threshold detector is integrated into network level GLRT to improve the performance. The detection criteria of oracle and clairvoyant detectors are also derived. Simulation results show that with careful design of the threshold detector, the overall detection performance of double-detector scheme would be better than the sign-GLRT proposed in [1] and close to oracle and clairvoyant detectors. Also, the proposed detector is applied to spectrum sensing and the results are near the well known energy detector which uses the real valued data while the proposed detector only uses the sign of the data.
Tasks
Published 2016-07-02
URL http://arxiv.org/abs/1607.00494v1
PDF http://arxiv.org/pdf/1607.00494v1.pdf
PWC https://paperswithcode.com/paper/double-detector-for-sparse-signal-detection
Repo
Framework
comments powered by Disqus