October 16, 2019

3321 words 16 mins read

Paper Group ANR 1158

Paper Group ANR 1158

Learning graphs from data: A signal representation perspective. A Theoretically Guaranteed Deep Optimization Framework for Robust Compressive Sensing MRI. A strong baseline for question relevancy ranking. Covariance Function Pre-Training with m-Kernels for Accelerated Bayesian Optimisation. Securing Distributed Gradient Descent in High Dimensional …

Learning graphs from data: A signal representation perspective

Title Learning graphs from data: A signal representation perspective
Authors Xiaowen Dong, Dorina Thanou, Michael Rabbat, Pascal Frossard
Abstract The construction of a meaningful graph topology plays a crucial role in the effective representation, processing, analysis and visualization of structured data. When a natural choice of the graph is not readily available from the data sets, it is thus desirable to infer or learn a graph topology from the data. In this tutorial overview, we survey solutions to the problem of graph learning, including classical viewpoints from statistics and physics, and more recent approaches that adopt a graph signal processing (GSP) perspective. We further emphasize the conceptual similarities and differences between classical and GSP-based graph inference methods, and highlight the potential advantage of the latter in a number of theoretical and practical scenarios. We conclude with several open issues and challenges that are keys to the design of future signal processing and machine learning algorithms for learning graphs from data.
Tasks
Published 2018-06-03
URL https://arxiv.org/abs/1806.00848v3
PDF https://arxiv.org/pdf/1806.00848v3.pdf
PWC https://paperswithcode.com/paper/learning-graphs-from-data-a-signal
Repo
Framework

A Theoretically Guaranteed Deep Optimization Framework for Robust Compressive Sensing MRI

Title A Theoretically Guaranteed Deep Optimization Framework for Robust Compressive Sensing MRI
Authors Risheng Liu, Yuxi Zhang, Shichao Cheng, Xin Fan, Zhongxuan Luo
Abstract Magnetic Resonance Imaging (MRI) is one of the most dynamic and safe imaging techniques available for clinical applications. However, the rather slow speed of MRI acquisitions limits the patient throughput and potential indi cations. Compressive Sensing (CS) has proven to be an efficient technique for accelerating MRI acquisition. The most widely used CS-MRI model, founded on the premise of reconstructing an image from an incompletely filled k-space, leads to an ill-posed inverse problem. In the past years, lots of efforts have been made to efficiently optimize the CS-MRI model. Inspired by deep learning techniques, some preliminary works have tried to incorporate deep architectures into CS-MRI process. Unfortunately, the convergence issues (due to the experience-based networks) and the robustness (i.e., lack real-world noise modeling) of these deeply trained optimization methods are still missing. In this work, we develop a new paradigm to integrate designed numerical solvers and the data-driven architectures for CS-MRI. By introducing an optimal condition checking mechanism, we can successfully prove the convergence of our established deep CS-MRI optimization scheme. Furthermore, we explicitly formulate the Rician noise distributions within our framework and obtain an extended CS-MRI network to handle the real-world nosies in the MRI process. Extensive experimental results verify that the proposed paradigm outperforms the existing state-of-the-art techniques both in reconstruction accuracy and efficiency as well as robustness to noises in real scene.
Tasks Compressive Sensing
Published 2018-11-09
URL http://arxiv.org/abs/1811.03782v3
PDF http://arxiv.org/pdf/1811.03782v3.pdf
PWC https://paperswithcode.com/paper/a-theoretically-guaranteed-deep-optimization
Repo
Framework

A strong baseline for question relevancy ranking

Title A strong baseline for question relevancy ranking
Authors Ana V. González-Garduño, Isabelle Augenstein, Anders Søgaard
Abstract The best systems at the SemEval-16 and SemEval-17 community question answering shared tasks – a task that amounts to question relevancy ranking – involve complex pipelines and manual feature engineering. Despite this, many of these still fail at beating the IR baseline, i.e., the rankings provided by Google’s search engine. We present a strong baseline for question relevancy ranking by training a simple multi-task feed forward network on a bag of 14 distance measures for the input question pair. This baseline model, which is fast to train and uses only language-independent features, outperforms the best shared task systems on the task of retrieving relevant previously asked questions.
Tasks Community Question Answering, Feature Engineering, Question Answering
Published 2018-08-27
URL http://arxiv.org/abs/1808.08836v1
PDF http://arxiv.org/pdf/1808.08836v1.pdf
PWC https://paperswithcode.com/paper/a-strong-baseline-for-question-relevancy
Repo
Framework

Covariance Function Pre-Training with m-Kernels for Accelerated Bayesian Optimisation

Title Covariance Function Pre-Training with m-Kernels for Accelerated Bayesian Optimisation
Authors Alistair Shilton, Sunil Gupta, Santu Rana, Pratibha Vellanki, Cheng Li, Laurence Park, Svetha Venkatesh, Alessandra Sutti, David Rubin, Thomas Dorin, Alireza Vahid, Murray Height
Abstract The paper presents a novel approach to direct covariance function learning for Bayesian optimisation, with particular emphasis on experimental design problems where an existing corpus of condensed knowledge is present. The method presented borrows techniques from reproducing kernel Banach space theory (specifically m-kernels) and leverages them to convert (or re-weight) existing covariance functions into new, problem-specific covariance functions. The key advantage of this approach is that rather than relying on the user to manually select (with some hyperparameter tuning and experimentation) an appropriate covariance function it constructs the covariance function to specifically match the problem at hand. The technique is demonstrated on two real-world problems - specifically alloy design and short-polymer fibre manufacturing - as well as a selected test function.
Tasks Bayesian Optimisation
Published 2018-02-15
URL http://arxiv.org/abs/1802.05370v2
PDF http://arxiv.org/pdf/1802.05370v2.pdf
PWC https://paperswithcode.com/paper/covariance-function-pre-training-with-m
Repo
Framework

Securing Distributed Gradient Descent in High Dimensional Statistical Learning

Title Securing Distributed Gradient Descent in High Dimensional Statistical Learning
Authors Lili Su, Jiaming Xu
Abstract We consider unreliable distributed learning systems wherein the training data is kept confidential by external workers, and the learner has to interact closely with those workers to train a model. In particular, we assume that there exists a system adversary that can adaptively compromise some workers; the compromised workers deviate from their local designed specifications by sending out arbitrarily malicious messages. We assume in each communication round, up to $q$ out of the $m$ workers suffer Byzantine faults. Each worker keeps a local sample of size $n$ and the total sample size is $N=nm$. We propose a secured variant of the gradient descent method that can tolerate up to a constant fraction of Byzantine workers, i.e., $q/m = O(1)$. Moreover, we show the statistical estimation error of the iterates converges in $O(\log N)$ rounds to $O(\sqrt{q/N} + \sqrt{d/N})$, where $d$ is the model dimension. As long as $q=O(d)$, our proposed algorithm achieves the optimal error rate $O(\sqrt{d/N})$. Our results are obtained under some technical assumptions. Specifically, we assume strongly-convex population risk. Nevertheless, the empirical risk (sample version) is allowed to be non-convex. The core of our method is to robustly aggregate the gradients computed by the workers based on the filtering procedure proposed by Steinhardt et al. On the technical front, deviating from the existing literature on robustly estimating a finite-dimensional mean vector, we establish a {\em uniform} concentration of the sample covariance matrix of gradients, and show that the aggregated gradient, as a function of model parameter, converges uniformly to the true gradient function. To get a near-optimal uniform concentration bound, we develop a new matrix concentration inequality, which might be of independent interest.
Tasks
Published 2018-04-26
URL https://arxiv.org/abs/1804.10140v3
PDF https://arxiv.org/pdf/1804.10140v3.pdf
PWC https://paperswithcode.com/paper/securing-distributed-machine-learning-in-high
Repo
Framework

Predicting Out-of-View Feature Points for Model-Based Camera Pose Estimation

Title Predicting Out-of-View Feature Points for Model-Based Camera Pose Estimation
Authors Oliver Moolan-Feroze, Andrew Calway
Abstract In this work we present a novel framework that uses deep learning to predict object feature points that are out-of-view in the input image. This system was developed with the application of model-based tracking in mind, particularly in the case of autonomous inspection robots, where only partial views of the object are available. Out-of-view prediction is enabled by applying scaling to the feature point labels during network training. This is combined with a recurrent neural network architecture designed to provide the final prediction layers with rich feature information from across the spatial extent of the input image. To show the versatility of these out-of-view predictions, we describe how to integrate them in both a particle filter tracker and an optimisation based tracker. To evaluate our work we compared our framework with one that predicts only points inside the image. We show that as the amount of the object in view decreases, being able to predict outside the image bounds adds robustness to the final pose estimation.
Tasks Pose Estimation
Published 2018-03-05
URL http://arxiv.org/abs/1803.01577v1
PDF http://arxiv.org/pdf/1803.01577v1.pdf
PWC https://paperswithcode.com/paper/predicting-out-of-view-feature-points-for
Repo
Framework
Title Position Bias Estimation for Unbiased Learning-to-Rank in eCommerce Search
Authors Grigor Aslanyan, Utkarsh Porwal
Abstract The Unbiased Learning-to-Rank framework has been recently proposed as a general approach to systematically remove biases, such as position bias, from learning-to-rank models. The method takes two steps - estimating click propensities and using them to train unbiased models. Most common methods proposed in the literature for estimating propensities involve some degree of intervention in the live search engine. An alternative approach proposed recently uses an Expectation Maximization (EM) algorithm to estimate propensities by using ranking features for estimating relevances. In this work we propose a novel method to directly estimate propensities which does not use any intervention in live search or rely on modeling relevance. Rather, we take advantage of the fact that the same query-document pair may naturally change ranks over time. This typically occurs for eCommerce search because of change of popularity of items over time, existence of time dependent ranking features, or addition or removal of items to the index (an item getting sold or a new item being listed). However, our method is general and can be applied to any search engine for which the rank of the same document may naturally change over time for the same query. We derive a simple likelihood function that depends on propensities only, and by maximizing the likelihood we are able to get estimates of the propensities. We apply this method to eBay search data to estimate click propensities for web and mobile search and compare these with estimates using the EM method. We also use simulated data to show that the method gives reliable estimates of the “true” simulated propensities. Finally, we train an unbiased learning-to-rank model for eBay search using the estimated propensities and show that it outperforms both baselines - one without position bias correction and one with position bias correction using the EM method.
Tasks Learning-To-Rank
Published 2018-12-21
URL https://arxiv.org/abs/1812.09338v3
PDF https://arxiv.org/pdf/1812.09338v3.pdf
PWC https://paperswithcode.com/paper/direct-estimation-of-position-bias-for
Repo
Framework

Efficient Video Understanding via Layered Multi Frame-Rate Analysis

Title Efficient Video Understanding via Layered Multi Frame-Rate Analysis
Authors Ziyao Tang, Yongxi Lu, Tara Javidi
Abstract One of the greatest challenges in the design of a real-time perception system for autonomous driving vehicles and drones is the conflicting requirement of safety (high prediction accuracy) and efficiency. Traditional approaches use a single frame rate for the entire system. Motivated by the observation that the lack of robustness against environmental factors is the major weakness of compact ConvNet architectures, we propose a dual frame-rate system that brings in the best of both worlds: A modulator stream that executes an expensive models robust to environmental factors at a low frame rate to extract slowly changing features describing the environment, and a prediction stream that executes a light-weight model at real-time to extract transient signals that describes particularities of the current frame. The advantage of our design is validated by our extensive empirical study, showing that our solution leads to consistent improvements using a variety of backbone architecture choice and input resolutions. These findings suggest multiple frame-rate systems as a promising direction in designing efficient perception for autonomous agents.
Tasks Autonomous Driving, Video Understanding
Published 2018-11-24
URL http://arxiv.org/abs/1811.09834v1
PDF http://arxiv.org/pdf/1811.09834v1.pdf
PWC https://paperswithcode.com/paper/efficient-video-understanding-via-layered
Repo
Framework

A Stochastic Large-scale Machine Learning Algorithm for Distributed Features and Observations

Title A Stochastic Large-scale Machine Learning Algorithm for Distributed Features and Observations
Authors Biyi Fang, Diego Klabjan
Abstract As the size of modern data sets exceeds the disk and memory capacities of a single computer, machine learning practitioners have resorted to parallel and distributed computing. Given that optimization is one of the pillars of machine learning and predictive modeling, distributed optimization methods have recently garnered ample attention, in particular when either observations or features are distributed, but not both. We propose a general stochastic algorithm where observations, features, and gradient components can be sampled in a double distributed setting, i.e., with both features and observations distributed. Very technical analyses establish convergence properties of the algorithm under different conditions on the learning rate (diminishing to zero or constant). Computational experiments in Spark demonstrate a superior performance of our algorithm versus a benchmark in early iterations of the algorithm, which is due to the stochastic components of the algorithm.
Tasks Distributed Optimization
Published 2018-03-29
URL https://arxiv.org/abs/1803.11287v2
PDF https://arxiv.org/pdf/1803.11287v2.pdf
PWC https://paperswithcode.com/paper/a-stochastic-large-scale-machine-learning
Repo
Framework

Top-N-Rank: A Scalable List-wise Ranking Method for Recommender Systems

Title Top-N-Rank: A Scalable List-wise Ranking Method for Recommender Systems
Authors Junjie Liang, Jinlong Hu, Shoubin Dong, Vasant Honavar
Abstract We propose Top-N-Rank, a novel family of list-wise Learning-to-Rank models for reliably recommending the N top-ranked items. The proposed models optimize a variant of the widely used discounted cumulative gain (DCG) objective function which differs from DCG in two important aspects: (i) It limits the evaluation of DCG only on the top N items in the ranked lists, thereby eliminating the impact of low-ranked items on the learned ranking function; and (ii) it incorporates weights that allow the model to leverage multiple types of implicit feedback with differing levels of reliability or trustworthiness. Because the resulting objective function is non-smooth and hence challenging to optimize, we consider two smooth approximations of the objective function, using the traditional sigmoid function and the rectified linear unit (ReLU). We propose a family of learning-to-rank algorithms (Top-N-Rank) that work with any smooth objective function. Then, a more efficient variant, Top-N-Rank.ReLU, is introduced, which effectively exploits the properties of ReLU function to reduce the computational complexity of Top-N-Rank from quadratic to linear in the average number of items rated by users. The results of our experiments using two widely used benchmarks, namely, the MovieLens data set and the Amazon Video Games data set demonstrate that: (i) The `top-N truncation’ of the objective function substantially improves the ranking quality of the top N recommendations; (ii) using the ReLU for smoothing the objective function yields significant improvement in both ranking quality as well as runtime as compared to using the sigmoid; and (iii) Top-N-Rank.ReLU substantially outperforms the well-performing list-wise ranking methods in terms of ranking quality. |
Tasks Learning-To-Rank, Recommendation Systems
Published 2018-12-10
URL http://arxiv.org/abs/1812.04109v2
PDF http://arxiv.org/pdf/1812.04109v2.pdf
PWC https://paperswithcode.com/paper/top-n-rank-a-scalable-list-wise-ranking
Repo
Framework

On the Brain Networks of Complex Problem Solving

Title On the Brain Networks of Complex Problem Solving
Authors Abdullah Alchihabi, Omer Ekmekci, Baran B. Kivilcim, Sharlene D. Newman, Fatos T. Yarman Vural
Abstract Complex problem solving is a high level cognitive process which has been thoroughly studied over the last decade. The Tower of London (TOL) is a task that has been widely used to study problem-solving. In this study, we aim to explore the underlying cognitive network dynamics among anatomical regions of complex problem solving and its sub-phases, namely planning and execution. A new brain network construction model establishing dynamic functional brain networks using fMRI is proposed. The first step of the model is a preprocessing pipeline that manages to decrease the spatial redundancy while increasing the temporal resolution of the fMRI recordings. Then, dynamic brain networks are estimated using artificial neural networks. The network properties of the estimated brain networks are studied in order to identify regions of interest, such as hubs and subgroups of densely connected brain regions. The major similarities and dissimilarities of the network structure of planning and execution phases are highlighted. Our findings show the hubs and clusters of densely interconnected regions during both subtasks. It is observed that there are more hubs during the planning phase compared to the execution phase, and the clusters are more strongly connected during planning compared to execution.
Tasks
Published 2018-10-10
URL http://arxiv.org/abs/1810.05077v1
PDF http://arxiv.org/pdf/1810.05077v1.pdf
PWC https://paperswithcode.com/paper/on-the-brain-networks-of-complex-problem
Repo
Framework

Dynamically Context-Sensitive Time-Decay Attention for Dialogue Modeling

Title Dynamically Context-Sensitive Time-Decay Attention for Dialogue Modeling
Authors Shang-Yu Su, Pei-Chieh Yuan, Yun-Nung Chen
Abstract Spoken language understanding (SLU) is an essential component in conversational systems. Considering that contexts provide informative cues for better understanding, history can be leveraged for contextual SLU. However, most prior work only paid attention to the related content in history utterances and ignored the temporal information. In dialogues, it is intuitive that the most recent utterances are more important than the least recent ones, and time-aware attention should be in a decaying manner. Therefore, this paper allows the model to automatically learn a time-decay attention function where the attentional weights can be dynamically decided based on the content of each role’s contexts, which effectively integrates both content-aware and time-aware perspectives and demonstrates remarkable flexibility to complex dialogue contexts. The experiments on the benchmark Dialogue State Tracking Challenge (DSTC4) dataset show that the proposed dynamically context-sensitive time-decay attention mechanisms significantly improve the state-of-the-art model for contextual understanding performance.
Tasks Dialogue State Tracking, Spoken Language Understanding
Published 2018-09-05
URL http://arxiv.org/abs/1809.01557v2
PDF http://arxiv.org/pdf/1809.01557v2.pdf
PWC https://paperswithcode.com/paper/dynamically-context-sensitive-time-decay
Repo
Framework

Intervention Harvesting for Context-Dependent Examination-Bias Estimation

Title Intervention Harvesting for Context-Dependent Examination-Bias Estimation
Authors Zhichong Fang, Aman Agarwal, Thorsten Joachims
Abstract Accurate estimates of examination bias are crucial for unbiased learning-to-rank from implicit feedback in search engines and recommender systems, since they enable the use of Inverse Propensity Score (IPS) weighting techniques to address selection biases and missing data. Unfortunately, existing examination-bias estimators are limited to the Position-Based Model (PBM), where the examination bias may only depend on the rank of the document. To overcome this limitation, we propose a Contextual Position-Based Model (CPBM) where the examination bias may also depend on a context vector describing the query and the user. Furthermore, we propose an effective estimator for the CPBM based on intervention harvesting. A key feature of the estimator is that it does not require disruptive interventions but merely exploits natural variation resulting from the use of multiple historic ranking functions. Real-world experiments on the ArXiv search engine and semi-synthetic experiments on the Yahoo Learning-To-Rank dataset demonstrate the superior effectiveness and robustness of the new approach.
Tasks Learning-To-Rank, Recommendation Systems
Published 2018-11-05
URL https://arxiv.org/abs/1811.01802v3
PDF https://arxiv.org/pdf/1811.01802v3.pdf
PWC https://paperswithcode.com/paper/intervention-harvesting-for-context-dependent
Repo
Framework

Consistent polynomial-time unseeded graph matching for Lipschitz graphons

Title Consistent polynomial-time unseeded graph matching for Lipschitz graphons
Authors Yuan Zhang
Abstract We propose a consistent polynomial-time method for the unseeded node matching problem for networks with smooth underlying structures. Despite widely conjectured by the research community that the structured graph matching problem to be significantly easier than its worst case counterpart, well-known to be NP-hard, the statistical version of the problem has stood a challenge that resisted any solution both provable and polynomial-time. The closest existing work requires quasi-polynomial time. Our method is based on the latest advances in graphon estimation techniques and analysis on the concentration of empirical Wasserstein distances. Its core is a simple yet unconventional sampling-and-matching scheme that reduces the problem from unseeded to seeded. Our method allows flexible efficiencies, is convenient to analyze and potentially can be extended to more general settings. Our work enables a rich variety of subsequent estimations and inferences.
Tasks Graph Matching, Graphon Estimation
Published 2018-07-29
URL http://arxiv.org/abs/1807.11027v1
PDF http://arxiv.org/pdf/1807.11027v1.pdf
PWC https://paperswithcode.com/paper/consistent-polynomial-time-unseeded-graph
Repo
Framework

Learning to Rank Query Graphs for Complex Question Answering over Knowledge Graphs

Title Learning to Rank Query Graphs for Complex Question Answering over Knowledge Graphs
Authors Gaurav Maheshwari, Priyansh Trivedi, Denis Lukovnikov, Nilesh Chakraborty, Asja Fischer, Jens Lehmann
Abstract In this paper, we conduct an empirical investigation of neural query graph ranking approaches for the task of complex question answering over knowledge graphs. We experiment with six different ranking models and propose a novel self-attention based slot matching model which exploits the inherent structure of query graphs, our logical form of choice. Our proposed model generally outperforms the other models on two QA datasets over the DBpedia knowledge graph, evaluated in different settings. In addition, we show that transfer learning from the larger of those QA datasets to the smaller dataset yields substantial improvements, effectively offsetting the general lack of training data.
Tasks Graph Ranking, Knowledge Graphs, Learning-To-Rank, Question Answering, Transfer Learning
Published 2018-11-02
URL http://arxiv.org/abs/1811.01118v1
PDF http://arxiv.org/pdf/1811.01118v1.pdf
PWC https://paperswithcode.com/paper/learning-to-rank-query-graphs-for-complex
Repo
Framework
comments powered by Disqus