January 27, 2020

2941 words 14 mins read

Paper Group ANR 1253

Paper Group ANR 1253

Efficient Bimanual Manipulation Using Learned Task Schemas. Exploiting Channel Similarity for Accelerating Deep Convolutional Neural Networks. Tensor Completion via Gaussian Process Based Initialization. J-MoDL: Joint Model-Based Deep Learning for Optimized Sampling and Reconstruction. Federated machine learning with Anonymous Random Hybridization …

Efficient Bimanual Manipulation Using Learned Task Schemas

Title Efficient Bimanual Manipulation Using Learned Task Schemas
Authors Rohan Chitnis, Shubham Tulsiani, Saurabh Gupta, Abhinav Gupta
Abstract We address the problem of effectively composing skills to solve sparse-reward tasks in the real world. Given a set of parameterized skills (such as exerting a force or doing a top grasp at a location), our goal is to learn policies that invoke these skills to efficiently solve such tasks. Our insight is that for many tasks, the learning process can be decomposed into learning a state-independent task schema (a sequence of skills to execute) and a policy to choose the parameterizations of the skills in a state-dependent manner. For such tasks, we show that explicitly modeling the schema’s state-independence can yield significant improvements in sample efficiency for model-free reinforcement learning algorithms. Furthermore, these schemas can be transferred to solve related tasks, by simply re-learning the parameterizations with which the skills are invoked. We find that doing so enables learning to solve sparse-reward tasks on real-world robotic systems very efficiently. We validate our approach experimentally over a suite of robotic bimanual manipulation tasks, both in simulation and on real hardware. See videos at http://tinyurl.com/chitnis-schema.
Tasks
Published 2019-09-30
URL https://arxiv.org/abs/1909.13874v2
PDF https://arxiv.org/pdf/1909.13874v2.pdf
PWC https://paperswithcode.com/paper/efficient-bimanual-manipulation-using-learned
Repo
Framework

Exploiting Channel Similarity for Accelerating Deep Convolutional Neural Networks

Title Exploiting Channel Similarity for Accelerating Deep Convolutional Neural Networks
Authors Yunxiang Zhang, Chenglong Zhao, Bingbing Ni, Jian Zhang, Haoran Deng
Abstract To address the limitations of existing magnitude-based pruning algorithms in cases where model weights or activations are of large and similar magnitude, we propose a novel perspective to discover parameter redundancy among channels and accelerate deep CNNs via channel pruning. Precisely, we argue that channels revealing similar feature information have functional overlap and that most channels within each such similarity group can be removed without compromising model’s representational power. After deriving an effective metric for evaluating channel similarity through probabilistic modeling, we introduce a pruning algorithm via hierarchical clustering of channels. In particular, the proposed algorithm does not rely on sparsity training techniques or complex data-driven optimization and can be directly applied to pre-trained models. Extensive experiments on benchmark datasets strongly demonstrate the superior acceleration performance of our approach over prior arts. On ImageNet, our pruned ResNet-50 with 30% FLOPs reduced outperforms the baseline model.
Tasks
Published 2019-08-06
URL https://arxiv.org/abs/1908.02620v1
PDF https://arxiv.org/pdf/1908.02620v1.pdf
PWC https://paperswithcode.com/paper/exploiting-channel-similarity-for
Repo
Framework

Tensor Completion via Gaussian Process Based Initialization

Title Tensor Completion via Gaussian Process Based Initialization
Authors Yermek Kapushev, Ivan Oseledets, Evgeny Burnaev
Abstract In this paper, we consider the tensor completion problem representing the solution in the tensor train (TT) format. It is assumed that tensor is high-dimensional, and tensor values are generated by an unknown smooth function. The assumption allows us to develop an efficient initialization scheme based on Gaussian Process Regression and TT-cross approximation technique. The proposed approach can be used in conjunction with any optimization algorithm that is usually utilized in tensor completion problems. We empirically justify that in this case the reconstruction error improves compared to the tensor completion with random initialization. As an additional benefit, our technique automatically selects rank thanks to using the TT-cross approximation technique.
Tasks
Published 2019-12-11
URL https://arxiv.org/abs/1912.05179v1
PDF https://arxiv.org/pdf/1912.05179v1.pdf
PWC https://paperswithcode.com/paper/tensor-completion-via-gaussian-process-based
Repo
Framework

J-MoDL: Joint Model-Based Deep Learning for Optimized Sampling and Reconstruction

Title J-MoDL: Joint Model-Based Deep Learning for Optimized Sampling and Reconstruction
Authors Hemant Kumar Aggarwal, Mathews Jacob
Abstract Modern MRI schemes, which rely on compressed sensing or deep learning algorithms to recover MRI data from undersampled multichannel Fourier measurements, are widely used to reduce scan time. The image quality of these approaches is heavily dependent on the sampling pattern. We introduce a continuous strategy to jointly optimize the sampling pattern and the parameters of the reconstruction algorithm. We propose to use a model-based deep learning (MoDL) image reconstruction algorithm, which alternates between a data consistency module and a convolutional neural network (CNN). We use a multi-channel forward model, consisting of a non-uniform Fourier transform with continuously defined sampling locations, to realize the data consistency block. This approach facilitates the joint and continuous optimization of the sampling pattern and the CNN parameters. We observe that the joint optimization of the sampling patterns and the reconstruction module significantly improves the performance, compared to current deep learning methods that use variable density sampling patterns. Our experiments show that the improved decoupling of the CNN parameters from the sampling scheme offered by the MoDL scheme translates to improved optimization and performance compared to a similar scheme using a direct-inversion based reconstruction algorithm. The experiments also show that the proposed scheme offers good convergence and reduces the dependence on initialization.
Tasks Image Reconstruction
Published 2019-11-06
URL https://arxiv.org/abs/1911.02945v2
PDF https://arxiv.org/pdf/1911.02945v2.pdf
PWC https://paperswithcode.com/paper/joint-optimization-of-sampling-patterns-and
Repo
Framework

Federated machine learning with Anonymous Random Hybridization (FeARH) on medical records

Title Federated machine learning with Anonymous Random Hybridization (FeARH) on medical records
Authors Jianfei Cui, Dianbo Liu
Abstract Sometimes electrical medical records are restricted and difficult to centralize for machine learning, which could only be trained in distributed manner that involved many institutions in the process. However, sometimes some institutions are likely to figure out the private data used for training certain models based on the parameters they obtained, which is a violation of privacy and certain regulations. Under those circumstances, we develop an algorithm, called ‘federated machine learning with anonymous random hybridization’(abbreviated as ‘FeARH’), using mainly hybridization algorithm to eliminate connections between medical record data and models’ parameters, which avoid untrustworthy institutions from stealing patients’ private medical records. Based on our experiment, our new algorithm has similar AUCROC and AUCPR result compared with machine learning in centralized manner and original federated machine learning, at the same time, our algorithm can greatly reduce data transfer size in comparison with original federated machine learning.
Tasks
Published 2019-12-25
URL https://arxiv.org/abs/2001.09751v1
PDF https://arxiv.org/pdf/2001.09751v1.pdf
PWC https://paperswithcode.com/paper/federated-machine-learning-with-anonymous
Repo
Framework

Using Monolingual Data in Neural Machine Translation: a Systematic Study

Title Using Monolingual Data in Neural Machine Translation: a Systematic Study
Authors Franck Burlot, François Yvon
Abstract Neural Machine Translation (MT) has radically changed the way systems are developed. A major difference with the previous generation (Phrase-Based MT) is the way monolingual target data, which often abounds, is used in these two paradigms. While Phrase-Based MT can seamlessly integrate very large language models trained on billions of sentences, the best option for Neural MT developers seems to be the generation of artificial parallel data through \textsl{back-translation} - a technique that fails to fully take advantage of existing datasets. In this paper, we conduct a systematic study of back-translation, comparing alternative uses of monolingual data, as well as multiple data generation procedures. Our findings confirm that back-translation is very effective and give new explanations as to why this is the case. We also introduce new data simulation techniques that are almost as effective, yet much cheaper to implement.
Tasks Machine Translation
Published 2019-03-27
URL http://arxiv.org/abs/1903.11437v1
PDF http://arxiv.org/pdf/1903.11437v1.pdf
PWC https://paperswithcode.com/paper/using-monolingual-data-in-neural-machine-1
Repo
Framework

Gray Level Image Threshold Using Neutrosophic Shannon Entropy

Title Gray Level Image Threshold Using Neutrosophic Shannon Entropy
Authors Vasile Patrascu
Abstract This article presents a new method of segmenting grayscale images by minimizing Shannon’s neutrosophic entropy. For the proposed segmentation method, the neutrosophic information components, i.e., the degree of truth, the degree of neutrality and the degree of falsity are defined taking into account the belonging to the segmented regions and at the same time to the separation threshold area. The principle of the method is simple and easy to understand and can lead to multiple thresholds. The efficacy of the method is illustrated using some test gray level images. The experimental results show that the proposed method has good performance for segmentation with optimal gray level thresholds.
Tasks
Published 2019-06-26
URL https://arxiv.org/abs/1906.12167v1
PDF https://arxiv.org/pdf/1906.12167v1.pdf
PWC https://paperswithcode.com/paper/gray-level-image-threshold-using-neutrosophic
Repo
Framework

Data Selection with Cluster-Based Language Difference Models and Cynical Selection

Title Data Selection with Cluster-Based Language Difference Models and Cynical Selection
Authors Lucía Santamaría, Amittai Axelrod
Abstract We present and apply two methods for addressing the problem of selecting relevant training data out of a general pool for use in tasks such as machine translation. Building on existing work on class-based language difference models, we first introduce a cluster-based method that uses Brown clusters to condense the vocabulary of the corpora. Secondly, we implement the cynical data selection method, which incrementally constructs a training corpus to efficiently model the task corpus. Both the cluster-based and the cynical data selection approaches are used for the first time within a machine translation system, and we perform a head-to-head comparison. Our intrinsic evaluations show that both new methods outperform the standard Moore-Lewis approach (cross-entropy difference), in terms of better perplexity and OOV rates on in-domain data. The cynical approach converges much quicker, covering nearly all of the in-domain vocabulary with 84% less data than the other methods. Furthermore, the new approaches can be used to select machine translation training data for training better systems. Our results confirm that class-based selection using Brown clusters is a viable alternative to POS-based class-based methods, and removes the reliance on a part-of-speech tagger. Additionally, we are able to validate the recently proposed cynical data selection method, showing that its performance in SMT models surpasses that of traditional cross-entropy difference methods and more closely matches the sentence length of the task corpus.
Tasks Machine Translation
Published 2019-04-09
URL http://arxiv.org/abs/1904.04900v1
PDF http://arxiv.org/pdf/1904.04900v1.pdf
PWC https://paperswithcode.com/paper/data-selection-with-cluster-based-language
Repo
Framework

A Rademacher Complexity Based Method fo rControlling Power and Confidence Level in Adaptive Statistical Analysis

Title A Rademacher Complexity Based Method fo rControlling Power and Confidence Level in Adaptive Statistical Analysis
Authors Lorenzo De Stefani, Eli Upfal
Abstract While standard statistical inference techniques and machine learning generalization bounds assume that tests are run on data selected independently of the hypotheses, practical data analysis and machine learning are usually iterative and adaptive processes where the same holdout data is often used for testing a sequence of hypotheses (or models), which may each depend on the outcome of the previous tests on the same data. In this work, we present RadaBound a rigorous, efficient and practical procedure for controlling the generalization error when using a holdout sample for multiple adaptive testing. Our solution is based on a new application of the Rademacher Complexity generalization bounds, adapted to dependent tests. We demonstrate the statistical power and practicality of our method through extensive simulations and comparisons to alternative approaches.
Tasks
Published 2019-10-04
URL https://arxiv.org/abs/1910.03493v1
PDF https://arxiv.org/pdf/1910.03493v1.pdf
PWC https://paperswithcode.com/paper/a-rademacher-complexity-based-method-fo
Repo
Framework

Learning Improved Representations by Transferring Incomplete Evidence Across Heterogeneous Tasks

Title Learning Improved Representations by Transferring Incomplete Evidence Across Heterogeneous Tasks
Authors Athanasios Davvetas, Iraklis A. Klampanos
Abstract Acquiring ground truth labels for unlabelled data can be a costly procedure, since it often requires manual labour that is error-prone. Consequently, the available amount of labelled data is increasingly reduced due to the limitations of manual data labelling. It is possible to increase the amount of labelled data samples by performing automated labelling or crowd-sourcing the annotation procedure. However, they often introduce noise or uncertainty in the labelset, that leads to decreased performance of supervised deep learning methods. On the other hand, weak supervision methods remain robust during noisy labelsets or can be effective even with low amounts of labelled data. In this paper we evaluate the effectiveness of a representation learning method that uses external categorical evidence called “Evidence Transfer”, against low amount of corresponding evidence termed as incomplete evidence. Evidence transfer is a robust solution against external unknown categorical evidence that can introduce noise or uncertainty. In our experimental evaluation, evidence transfer proves to be effective and robust against different levels of incompleteness, for two types of incomplete evidence.
Tasks Representation Learning
Published 2019-12-22
URL https://arxiv.org/abs/1912.10490v1
PDF https://arxiv.org/pdf/1912.10490v1.pdf
PWC https://paperswithcode.com/paper/learning-improved-representations-by
Repo
Framework

Deformable Filter Convolution for Point Cloud Reasoning

Title Deformable Filter Convolution for Point Cloud Reasoning
Authors Yuwen Xiong, Mengye Ren, Renjie Liao, Kelvin Wong, Raquel Urtasun
Abstract Point clouds are the native output of many real-world 3D sensors. To borrow the success of 2D convolutional network architectures, a majority of popular 3D perception models voxelize the points, which can result in a loss of local geometric details that cannot be recovered. In this paper, we propose a novel learnable convolution layer for processing 3D point cloud data directly. Instead of discretizing points into fixed voxels, we deform our learnable 3D filters to match with the point cloud shape. We propose to combine voxelized backbone networks with our deformable filter layer at 1) the network input stream and 2) the output prediction layers to enhance point level reasoning. We obtain state-of-the-art results on LiDAR semantic segmentation and producing a significant gain in performance on LiDAR object detection.
Tasks Object Detection, Semantic Segmentation
Published 2019-07-30
URL https://arxiv.org/abs/1907.13079v1
PDF https://arxiv.org/pdf/1907.13079v1.pdf
PWC https://paperswithcode.com/paper/deformable-filter-convolution-for-point-cloud
Repo
Framework

JGAN: A Joint Formulation of GAN for Synthesizing Images and Labels

Title JGAN: A Joint Formulation of GAN for Synthesizing Images and Labels
Authors Minje Park
Abstract Image generation with explicit condition or label generally works better than unconditional image generation. In modern GAN frameworks, both generator and discriminator are formulated to model the conditional distribution of images given with labels. In this paper, we provide an alternative formulation of GAN which models joint distribution of images and labels. There are two advantages in this joint formulation over conditional approaches. The first advantage is that the joint formulation is more robust to label noises, and the second is we can use any kind of weak labels (or additional information which has dependence on the original image data) to enhance unconditional image generation. We will show the effectiveness of joint formulation in CIFAR-10, CIFAR-100, and STL dataset.
Tasks Image Generation
Published 2019-05-28
URL https://arxiv.org/abs/1905.11574v2
PDF https://arxiv.org/pdf/1905.11574v2.pdf
PWC https://paperswithcode.com/paper/jgan-a-joint-formulation-of-gan-for
Repo
Framework

Maximum Weighted Loss Discrepancy

Title Maximum Weighted Loss Discrepancy
Authors Fereshte Khani, Aditi Raghunathan, Percy Liang
Abstract Though machine learning algorithms excel at minimizing the average loss over a population, this might lead to large discrepancies between the losses across groups within the population. To capture this inequality, we introduce and study a notion we call maximum weighted loss discrepancy (MWLD), the maximum (weighted) difference between the loss of a group and the loss of the population. We relate MWLD to group fairness notions and robustness to demographic shifts. We then show MWLD satisfies the following three properties: 1) It is statistically impossible to estimate MWLD when all groups have equal weights. 2) For a particular family of weighting functions, we can estimate MWLD efficiently. 3) MWLD is related to loss variance, a quantity that arises in generalization bounds. We estimate MWLD with different weighting functions on four common datasets from the fairness literature. We finally show that loss variance regularization can halve the loss variance of a classifier and hence reduce MWLD without suffering a significant drop in accuracy.
Tasks
Published 2019-06-08
URL https://arxiv.org/abs/1906.03518v1
PDF https://arxiv.org/pdf/1906.03518v1.pdf
PWC https://paperswithcode.com/paper/maximum-weighted-loss-discrepancy
Repo
Framework

On the Adaptivity of Stochastic Gradient-Based Optimization

Title On the Adaptivity of Stochastic Gradient-Based Optimization
Authors Lihua Lei, Michael I. Jordan
Abstract Stochastic-gradient-based optimization has been a core enabling methodology in applications to large-scale problems in machine learning and related areas. Despite the progress, the gap between theory and practice remains significant, with theoreticians pursuing mathematical optimality at a cost of obtaining specialized procedures in different regimes (e.g., modulus of strong convexity, magnitude of target accuracy, signal-to-noise ratio), and with practitioners not readily able to know which regime is appropriate to their problem, and seeking broadly applicable algorithms that are reasonably close to optimality. To bridge these perspectives it is necessary to study algorithms that are adaptive to different regimes. We present the stochastically controlled stochastic gradient (SCSG) method for composite convex finite-sum optimization problems and show that SCSG is adaptive to both strong convexity and target accuracy. The adaptivity is achieved by batch variance reduction with adaptive batch sizes and a novel technique, which we referred to as \emph{geometrization}, which sets the length of each epoch as a geometric random variable. The algorithm achieves strictly better theoretical complexity than other existing adaptive algorithms, while the tuning parameters of the algorithm only depend on the smoothness parameter of the objective.
Tasks
Published 2019-04-09
URL https://arxiv.org/abs/1904.04480v2
PDF https://arxiv.org/pdf/1904.04480v2.pdf
PWC https://paperswithcode.com/paper/on-the-adaptivity-of-stochastic-gradient
Repo
Framework

SPEAK YOUR MIND! Towards Imagined Speech Recognition With Hierarchical Deep Learning

Title SPEAK YOUR MIND! Towards Imagined Speech Recognition With Hierarchical Deep Learning
Authors Pramit Saha, Muhammad Abdul-Mageed, Sidney Fels
Abstract Speech-related Brain Computer Interface (BCI) technologies provide effective vocal communication strategies for controlling devices through speech commands interpreted from brain signals. In order to infer imagined speech from active thoughts, we propose a novel hierarchical deep learning BCI system for subject-independent classification of 11 speech tokens including phonemes and words. Our novel approach exploits predicted articulatory information of six phonological categories (e.g., nasal, bilabial) as an intermediate step for classifying the phonemes and words, thereby finding discriminative signal responsible for natural speech synthesis. The proposed network is composed of hierarchical combination of spatial and temporal CNN cascaded with a deep autoencoder. Our best models on the KARA database achieve an average accuracy of 83.42% across the six different binary phonological classification tasks, and 53.36% for the individual token identification task, significantly outperforming our baselines. Ultimately, our work suggests the possible existence of a brain imagery footprint for the underlying articulatory movement related to different sounds that can be used to aid imagined speech decoding.
Tasks Speech Recognition, Speech Synthesis
Published 2019-04-08
URL http://arxiv.org/abs/1904.05746v1
PDF http://arxiv.org/pdf/1904.05746v1.pdf
PWC https://paperswithcode.com/paper/speak-your-mind-towards-imagined-speech
Repo
Framework
comments powered by Disqus