January 28, 2020

3048 words 15 mins read

Paper Group ANR 781

Paper Group ANR 781

Bayesian Network Based Label Correlation Analysis For Multi-label Classifier Chain. An Automatic Digital Terrain Generation Technique for Terrestrial Sensing and Virtual Reality Applications. Modelling Dynamic Interactions between Relevance Dimensions. Cognitive Model Priors for Predicting Human Decisions. Unremarkable AI: Fitting Intelligent Decis …

Bayesian Network Based Label Correlation Analysis For Multi-label Classifier Chain

Title Bayesian Network Based Label Correlation Analysis For Multi-label Classifier Chain
Authors Ran Wang, Suhe Ye, Ke Li, Sam Kwong
Abstract Classifier chain (CC) is a multi-label learning approach that constructs a sequence of binary classifiers according to a label order. Each classifier in the sequence is responsible for predicting the relevance of one label. When training the classifier for a label, proceeding labels will be taken as extended features. If the extended features are highly correlated to the label, the performance will be improved, otherwise, the performance will not be influenced or even degraded. How to discover label correlation and determine the label order is critical for CC approach. This paper employs Bayesian network (BN) to model the label correlations and proposes a new BN-based CC method (BNCC). First, conditional entropy is used to describe the dependency relations among labels. Then, a BN is built up by taking nodes as labels and weights of edges as their dependency relations. A new scoring function is proposed to evaluate a BN structure, and a heuristic algorithm is introduced to optimize the BN. At last, by applying topological sorting on the nodes of the optimized BN, the label order for constructing CC model is derived. Experimental comparisons demonstrate the feasibility and effectiveness of the proposed method.
Tasks Multi-Label Learning
Published 2019-08-06
URL https://arxiv.org/abs/1908.02172v1
PDF https://arxiv.org/pdf/1908.02172v1.pdf
PWC https://paperswithcode.com/paper/bayesian-network-based-label-correlation
Repo
Framework

An Automatic Digital Terrain Generation Technique for Terrestrial Sensing and Virtual Reality Applications

Title An Automatic Digital Terrain Generation Technique for Terrestrial Sensing and Virtual Reality Applications
Authors Lee Easson, Alireza Tavakkoli, Jonathan Greenberg
Abstract The identification and modeling of the terrain from point cloud data is an important component of Terrestrial Remote Sensing (TRS) applications. The main focus in terrain modeling is capturing details of complex geological features of landforms. Traditional terrain modeling approaches rely on the user to exert control over terrain features. However, relying on the user input to manually develop the digital terrain becomes intractable when considering the amount of data generated by new remote sensing systems capable of producing massive aerial and ground-based point clouds from scanned environments. This article provides a novel terrain modeling technique capable of automatically generating accurate and physically realistic Digital Terrain Models (DTM) from a variety of point cloud data. The proposed method runs efficiently on large-scale point cloud data with real-time performance over large segments of terrestrial landforms. Moreover, generated digital models are designed to effectively render within a Virtual Reality (VR) environment in real time. The paper concludes with an in-depth discussion of possible research directions and outstanding technical and scientific challenges to improve the proposed approach.
Tasks
Published 2019-10-11
URL https://arxiv.org/abs/1910.04944v1
PDF https://arxiv.org/pdf/1910.04944v1.pdf
PWC https://paperswithcode.com/paper/an-automatic-digital-terrain-generation
Repo
Framework

Modelling Dynamic Interactions between Relevance Dimensions

Title Modelling Dynamic Interactions between Relevance Dimensions
Authors Sagar Uprety, Shahram Dehdashti, Lauren Fell, Peter Bruza, Dawei Song
Abstract Relevance is an underlying concept in the field of Information Science and Retrieval. It is a cognitive notion consisting of several different criteria or dimensions. Theoretical models of relevance allude to interdependence between these dimensions, where their interaction and fusion leads to the final inference of relevance. We study the interaction between the relevance dimensions using the mathematical framework of Quantum Theory. It is considered a generalised framework to model decision making under uncertainty, involving multiple perspectives and influenced by context. Specifically, we conduct a user study by constructing the cognitive analogue of a famous experiment in Quantum Physics. The data is used to construct a complex-valued vector space model of the user’s cognitive state, which is used to explain incompatibility and interference between relevance dimensions. The implications of our findings to inform the design of Information Retrieval systems are also discussed.
Tasks Decision Making, Decision Making Under Uncertainty, Information Retrieval
Published 2019-07-25
URL https://arxiv.org/abs/1907.10943v1
PDF https://arxiv.org/pdf/1907.10943v1.pdf
PWC https://paperswithcode.com/paper/modelling-dynamic-interactions-between
Repo
Framework

Cognitive Model Priors for Predicting Human Decisions

Title Cognitive Model Priors for Predicting Human Decisions
Authors David D. Bourgin, Joshua C. Peterson, Daniel Reichman, Thomas L. Griffiths, Stuart J. Russell
Abstract Human decision-making underlies all economic behavior. For the past four decades, human decision-making under uncertainty has continued to be explained by theoretical models based on prospect theory, a framework that was awarded the Nobel Prize in Economic Sciences. However, theoretical models of this kind have developed slowly, and robust, high-precision predictive models of human decisions remain a challenge. While machine learning is a natural candidate for solving these problems, it is currently unclear to what extent it can improve predictions obtained by current theories. We argue that this is mainly due to data scarcity, since noisy human behavior requires massive sample sizes to be accurately captured by off-the-shelf machine learning methods. To solve this problem, what is needed are machine learning models with appropriate inductive biases for capturing human behavior, and larger datasets. We offer two contributions towards this end: first, we construct “cognitive model priors” by pretraining neural networks with synthetic data generated by cognitive models (i.e., theoretical models developed by cognitive psychologists). We find that fine-tuning these networks on small datasets of real human decisions results in unprecedented state-of-the-art improvements on two benchmark datasets. Second, we present the first large-scale dataset for human decision-making, containing over 240,000 human judgments across over 13,000 decision problems. This dataset reveals the circumstances where cognitive model priors are useful, and provides a new standard for benchmarking prediction of human decisions under uncertainty.
Tasks Decision Making, Decision Making Under Uncertainty
Published 2019-05-22
URL https://arxiv.org/abs/1905.09397v1
PDF https://arxiv.org/pdf/1905.09397v1.pdf
PWC https://paperswithcode.com/paper/cognitive-model-priors-for-predicting-human
Repo
Framework

Unremarkable AI: Fitting Intelligent Decision Support into Critical, Clinical Decision-Making Processes

Title Unremarkable AI: Fitting Intelligent Decision Support into Critical, Clinical Decision-Making Processes
Authors Qian Yang, Aaron Steinfeld, John Zimmerman
Abstract Clinical decision support tools (DST) promise improved healthcare outcomes by offering data-driven insights. While effective in lab settings, almost all DSTs have failed in practice. Empirical research diagnosed poor contextual fit as the cause. This paper describes the design and field evaluation of a radically new form of DST. It automatically generates slides for clinicians’ decision meetings with subtly embedded machine prognostics. This design took inspiration from the notion of “Unremarkable Computing”, that by augmenting the users’ routines technology/AI can have significant importance for the users yet remain unobtrusive. Our field evaluation suggests clinicians are more likely to encounter and embrace such a DST. Drawing on their responses, we discuss the importance and intricacies of finding the right level of unremarkableness in DST design, and share lessons learned in prototyping critical AI systems as a situated experience.
Tasks Decision Making
Published 2019-04-21
URL http://arxiv.org/abs/1904.09612v1
PDF http://arxiv.org/pdf/1904.09612v1.pdf
PWC https://paperswithcode.com/paper/unremarkable-ai-fitting-intelligent-decision
Repo
Framework

A Distributionally Robust Boosting Algorithm

Title A Distributionally Robust Boosting Algorithm
Authors Jose Blanchet, Yang Kang, Fan Zhang, Zhangyi Hu
Abstract Distributionally Robust Optimization (DRO) has been shown to provide a flexible framework for decision making under uncertainty and statistical estimation. For example, recent works in DRO have shown that popular statistical estimators can be interpreted as the solutions of suitable formulated data-driven DRO problems. In turn, this connection is used to optimally select tuning parameters in terms of a principled approach informed by robustness considerations. This paper contributes to this growing literature, connecting DRO and statistics, by showing how boosting algorithms can be studied via DRO. We propose a boosting type algorithm, named DRO-Boosting, as a procedure to solve our DRO formulation. Our DRO-Boosting algorithm recovers Adaptive Boosting (AdaBoost) in particular, thus showing that AdaBoost is effectively solving a DRO problem. We apply our algorithm to a financial dataset on credit card default payment prediction. We find that our approach compares favorably to alternative boosting methods which are widely used in practice.
Tasks Decision Making, Decision Making Under Uncertainty
Published 2019-05-20
URL https://arxiv.org/abs/1905.07845v1
PDF https://arxiv.org/pdf/1905.07845v1.pdf
PWC https://paperswithcode.com/paper/a-distributionally-robust-boosting-algorithm
Repo
Framework

A Survey of Methods to Leverage Monolingual Data in Low-resource Neural Machine Translation

Title A Survey of Methods to Leverage Monolingual Data in Low-resource Neural Machine Translation
Authors Ilshat Gibadullin, Aidar Valeev, Albina Khusainova, Adil Khan
Abstract Neural machine translation has become the state-of-the-art for language pairs with large parallel corpora. However, the quality of machine translation for low-resource languages leaves much to be desired. There are several approaches to mitigate this problem, such as transfer learning, semi-supervised and unsupervised learning techniques. In this paper, we review the existing methods, where the main idea is to exploit the power of monolingual data, which, compared to parallel, is usually easier to obtain and significantly greater in amount.
Tasks Low-Resource Neural Machine Translation, Machine Translation, Transfer Learning
Published 2019-10-01
URL https://arxiv.org/abs/1910.00373v1
PDF https://arxiv.org/pdf/1910.00373v1.pdf
PWC https://paperswithcode.com/paper/a-survey-of-methods-to-leverage-monolingual
Repo
Framework

Learning Attribute Patterns in High-Dimensional Structured Latent Attribute Models

Title Learning Attribute Patterns in High-Dimensional Structured Latent Attribute Models
Authors Yuqi Gu, Gongjun Xu
Abstract Structured latent attribute models (SLAMs) are a special family of discrete latent variable models widely used in social and biological sciences. This paper considers the problem of learning significant attribute patterns from a SLAM with potentially high-dimensional configurations of the latent attributes. We address the theoretical identifiability issue, propose a penalized likelihood method for the selection of the attribute patterns, and further establish the selection consistency in such an overfitted SLAM with diverging number of latent patterns. The good performance of the proposed methodology is illustrated by simulation studies and two real datasets in educational assessment.
Tasks Latent Variable Models
Published 2019-04-08
URL https://arxiv.org/abs/1904.04378v2
PDF https://arxiv.org/pdf/1904.04378v2.pdf
PWC https://paperswithcode.com/paper/learning-attribute-patterns-in-high
Repo
Framework

Efficient Image Splicing Localization via Contrastive Feature Extraction

Title Efficient Image Splicing Localization via Contrastive Feature Extraction
Authors Ronald Salloum, C. -C. Jay Kuo
Abstract In this work, we propose a new data visualization and clustering technique for discovering discriminative structures in high-dimensional data. This technique, referred to as cPCA++, utilizes the fact that the interesting features of a “target” dataset may be obscured by high variance components during traditional PCA. By analyzing what is referred to as a “background” dataset (i.e., one that exhibits the high variance principal components but not the interesting structures), our technique is capable of efficiently highlighting the structure that is unique to the “target” dataset. Similar to another recently proposed algorithm called “contrastive PCA” (cPCA), the proposed cPCA++ method identifies important dataset specific patterns that are not detected by traditional PCA in a wide variety of settings. However, the proposed cPCA++ method is significantly more efficient than cPCA, because it does not require the parameter sweep in the latter approach. We applied the cPCA++ method to the problem of image splicing localization. In this application, we utilize authentic edges as the background dataset and the spliced edges as the target dataset. The proposed method is significantly more efficient than state-of-the-art methods, as the former does not require iterative updates of filter weights via stochastic gradient descent and backpropagation, nor the training of a classifier. Furthermore, the cPCA++ method is shown to provide performance scores comparable to the state-of-the-art Multi-task Fully Convolutional Network (MFCN).
Tasks
Published 2019-01-22
URL http://arxiv.org/abs/1901.07172v1
PDF http://arxiv.org/pdf/1901.07172v1.pdf
PWC https://paperswithcode.com/paper/efficient-image-splicing-localization-via
Repo
Framework

Online Evaluation of Audiences for Targeted Advertising via Bandit Experiments

Title Online Evaluation of Audiences for Targeted Advertising via Bandit Experiments
Authors Tong Geng, Xiliang Lin, Harikesh S. Nair
Abstract Firms implementing digital advertising campaigns face a complex problem in determining the right match between their advertising creatives and target audiences. Typical solutions to the problem have leveraged non-experimental methods, or used “split-testing” strategies that have not explicitly addressed the complexities induced by targeted audiences that can potentially overlap with one another. This paper presents an adaptive algorithm that addresses the problem via online experimentation. The algorithm is set up as a contextual bandit and addresses the overlap issue by partitioning the target audiences into disjoint, non-overlapping sub-populations. It learns an optimal creative display policy in the disjoint space, while assessing in parallel which creative has the best match in the space of possibly overlapping target audiences. Experiments show that the proposed method is more efficient compared to naive “split-testing” or non-adaptive “A/B/n” testing based methods. We also describe a testing product we built that uses the algorithm. The product is currently deployed on the advertising platform of JD.com, an eCommerce company and a publisher of digital ads in China.
Tasks
Published 2019-07-04
URL https://arxiv.org/abs/1907.02178v3
PDF https://arxiv.org/pdf/1907.02178v3.pdf
PWC https://paperswithcode.com/paper/online-evaluation-of-audiences-for-targeted
Repo
Framework

Discovering Common Change-Point Patterns in Functional Connectivity Across Subjects

Title Discovering Common Change-Point Patterns in Functional Connectivity Across Subjects
Authors Mengyu Dai, Zhengwu Zhang, Anuj Srivastava
Abstract This paper studies change-points in human brain functional connectivity (FC) and seeks patterns that are common across multiple subjects under identical external stimulus. FC relates to the similarity of fMRI responses across different brain regions when the brain is simply resting or performing a task. While the dynamic nature of FC is well accepted, this paper develops a formal statistical test for finding {\it change-points} in times series associated with FC. It represents short-term connectivity by a symmetric positive-definite matrix, and uses a Riemannian metric on this space to develop a graphical method for detecting change-points in a time series of such matrices. It also provides a graphical representation of estimated FC for stationary subintervals in between the detected change-points. Furthermore, it uses a temporal alignment of the test statistic, viewed as a real-valued function over time, to remove inter-subject variability and to discover common change-point patterns across subjects. This method is illustrated using data from Human Connectome Project (HCP) database for multiple subjects and tasks.
Tasks Time Series
Published 2019-04-26
URL http://arxiv.org/abs/1904.12023v1
PDF http://arxiv.org/pdf/1904.12023v1.pdf
PWC https://paperswithcode.com/paper/discovering-common-change-point-patterns-in
Repo
Framework

Entropy-Enhanced Multimodal Attention Model for Scene-Aware Dialogue Generation

Title Entropy-Enhanced Multimodal Attention Model for Scene-Aware Dialogue Generation
Authors Kuan-Yen Lin, Chao-Chun Hsu, Yun-Nung Chen, Lun-Wei Ku
Abstract With increasing information from social media, there are more and more videos available. Therefore, the ability to reason on a video is important and deserves to be discussed. TheDialog System Technology Challenge (DSTC7) (Yoshino et al. 2018) proposed an Audio Visual Scene-aware Dialog (AVSD) task, which contains five modalities including video, dialogue history, summary, and caption, as a scene-aware environment. In this paper, we propose the entropy-enhanced dynamic memory network (DMN) to effectively model video modality. The attention-based GRU in the proposed model can improve the model’s ability to comprehend and memorize sequential information. The entropy mechanism can control the attention distribution higher, so each to-be-answered question can focus more specifically on a small set of video segments. After the entropy-enhanced DMN secures the video context, we apply an attention model that in-corporates summary and caption to generate an accurate answer given the question about the video. In the official evaluation, our system can achieve improved performance against the released baseline model for both subjective and objective evaluation metrics.
Tasks Dialogue Generation, Scene-Aware Dialogue
Published 2019-08-22
URL https://arxiv.org/abs/1908.08191v1
PDF https://arxiv.org/pdf/1908.08191v1.pdf
PWC https://paperswithcode.com/paper/entropy-enhanced-multimodal-attention-model
Repo
Framework

Comment on “Solving Statistical Mechanics Using VANs”: Introducing saVANt - VANs Enhanced by Importance and MCMC Sampling

Title Comment on “Solving Statistical Mechanics Using VANs”: Introducing saVANt - VANs Enhanced by Importance and MCMC Sampling
Authors Kim Nicoli, Pan Kessel, Nils Strodthoff, Wojciech Samek, Klaus-Robert Müller, Shinichi Nakajima
Abstract In this comment on “Solving Statistical Mechanics Using Variational Autoregressive Networks” by Wu et al., we propose a subtle yet powerful modification of their approach. We show that the inherent sampling error of their method can be corrected by using neural network-based MCMC or importance sampling which leads to asymptotically unbiased estimators for physical quantities. This modification is possible due to a singular property of VANs, namely that they provide the exact sample probability. With these modifications, we believe that their method could have a substantially greater impact on various important fields of physics, including strongly-interacting field theories and statistical physics.
Tasks
Published 2019-03-26
URL http://arxiv.org/abs/1903.11048v1
PDF http://arxiv.org/pdf/1903.11048v1.pdf
PWC https://paperswithcode.com/paper/comment-on-solving-statistical-mechanics
Repo
Framework

NEARBY Platform for Detecting Asteroids in Astronomical Images Using Cloud-based Containerized Applications

Title NEARBY Platform for Detecting Asteroids in Astronomical Images Using Cloud-based Containerized Applications
Authors V. Bacu, A. Sabou, T. Stefanut, D. Gorgan, O. Vaduvescu
Abstract The continuing monitoring and surveying of the nearby space to detect Near Earth Objects (NEOs) and Near Earth Asteroids (NEAs) are essential because of the threats that this kind of objects impose on the future of our planet. We need more computational resources and advanced algorithms to deal with the exponential growth of the digital cameras’ performances and to be able to process (in near real-time) data coming from large surveys. This paper presents a software platform called NEARBY that supports automated detection of moving sources (asteroids) among stars from astronomical images. The detection procedure is based on the classic “blink” detection and, after that, the system supports visual analysis techniques to validate the moving sources, assisted by static and dynamical presentations.
Tasks
Published 2019-01-14
URL http://arxiv.org/abs/1901.04248v1
PDF http://arxiv.org/pdf/1901.04248v1.pdf
PWC https://paperswithcode.com/paper/nearby-platform-for-detecting-asteroids-in
Repo
Framework

Nonlocal Patches based Gaussian Mixture Model for Image Inpainting

Title Nonlocal Patches based Gaussian Mixture Model for Image Inpainting
Authors Wei Wan, Jun Liu
Abstract We consider the inpainting problem for noisy images. It is very challenge to suppress noise when image inpainting is processed. An image patches based nonlocal variational method is proposed to simultaneously inpainting and denoising in this paper. Our approach is developed on an assumption that the small image patches should be obeyed a distribution which can be described by a high dimension Gaussian Mixture Model. By a maximum a posteriori (MAP) estimation, we formulate a new regularization term according to the log-likelihood function of the mixture model. To optimize this regularization term efficiently, we adopt the idea of the Expectation Maximum (EM) algorithm. In which, the expectation step can give an adaptive weighting function which can be regarded as a nonlocal connections among pixels. Using this fact, we built a framework for non-local image inpainting under noise. Moreover, we mathematically prove the existence of minimizer for the proposed inpainting model. By using a spitting algorithm, the proposed model are able to realize image inpainting and denoising simultaneously. Numerical results show that the proposed method can produce impressive reconstructed results when the inpainting region is rather large.
Tasks Denoising, Image Inpainting
Published 2019-09-22
URL https://arxiv.org/abs/1909.09932v1
PDF https://arxiv.org/pdf/1909.09932v1.pdf
PWC https://paperswithcode.com/paper/190909932
Repo
Framework
comments powered by Disqus