January 30, 2020

3230 words 16 mins read

Paper Group ANR 212

Paper Group ANR 212

Learning fashion compatibility across apparel categories for outfit recommendation. Hybrid Approaches for our Participation to the n2c2 Challenge on Cohort Selection for Clinical Trials. The What-If Tool: Interactive Probing of Machine Learning Models. Deep Trajectory for Recognition of Human Behaviours. On Approximate Nonlinear Gaussian Message Pa …

Learning fashion compatibility across apparel categories for outfit recommendation

Title Learning fashion compatibility across apparel categories for outfit recommendation
Authors Luisa F. Polania, Satyajit Gupte
Abstract This paper addresses the problem of generating recommendations for completing the outfit given that a user is interested in a particular apparel item. The proposed method is based on a siamese network used for feature extraction followed by a fully-connected network used for learning a fashion compatibility metric. The embeddings generated by the siamese network are augmented with color histogram features motivated by the important role that color plays in determining fashion compatibility. The training of the network is formulated as a maximum a posteriori (MAP) problem where Laplacian distributions are assumed for the filters of the siamese network to promote sparsity and matrix-variate normal distributions are assumed for the weights of the metric network to efficiently exploit correlations between the input units of each fully-connected layer.
Tasks
Published 2019-05-01
URL http://arxiv.org/abs/1905.03703v1
PDF http://arxiv.org/pdf/1905.03703v1.pdf
PWC https://paperswithcode.com/paper/190503703
Repo
Framework

Hybrid Approaches for our Participation to the n2c2 Challenge on Cohort Selection for Clinical Trials

Title Hybrid Approaches for our Participation to the n2c2 Challenge on Cohort Selection for Clinical Trials
Authors Xavier Tannier, Nicolas Paris, Hugo Cisneros, Christel Daniel, Matthieu Doutreligne, Catherine Duclos, Nicolas Griffon, Claire Hassen-Khodja, Ivan Lerner, Adrien Parrot, Éric Sadou, Cyril Saussol, Pascal Vaillant
Abstract Objective: Natural language processing can help minimize human intervention in identifying patients meeting eligibility criteria for clinical trials, but there is still a long way to go to obtain a general and systematic approach that is useful for researchers. We describe two methods taking a step in this direction and present their results obtained during the n2c2 challenge on cohort selection for clinical trials. Materials and Methods: The first method is a weakly supervised method using an unlabeled corpus (MIMIC) to build a silver standard, by producing semi-automatically a small and very precise set of rules to detect some samples of positive and negative patients. This silver standard is then used to train a traditional supervised model. The second method is a terminology-based approach where a medical expert selects the appropriate concepts, and a procedure is defined to search the terms and check the structural or temporal constraints. Results: On the n2c2 dataset containing annotated data about 13 selection criteria on 288 patients, we obtained an overall F1-measure of 0.8969, which is the third best result out of 45 participant teams, with no statistically significant difference with the best-ranked team. Discussion: Both approaches obtained very encouraging results and apply to different types of criteria. The weakly supervised method requires explicit descriptions of positive and negative examples in some reports. The terminology-based method is very efficient when medical concepts carry most of the relevant information. Conclusion: It is unlikely that much more annotated data will be soon available for the task of identifying a wide range of patient phenotypes. One must focus on weakly or non-supervised learning methods using both structured and unstructured data and relying on a comprehensive representation of the patients.
Tasks
Published 2019-03-19
URL http://arxiv.org/abs/1903.07879v1
PDF http://arxiv.org/pdf/1903.07879v1.pdf
PWC https://paperswithcode.com/paper/hybrid-approaches-for-our-participation-to
Repo
Framework

The What-If Tool: Interactive Probing of Machine Learning Models

Title The What-If Tool: Interactive Probing of Machine Learning Models
Authors James Wexler, Mahima Pushkarna, Tolga Bolukbasi, Martin Wattenberg, Fernanda Viegas, Jimbo Wilson
Abstract A key challenge in developing and deploying Machine Learning (ML) systems is understanding their performance across a wide range of inputs. To address this challenge, we created the What-If Tool, an open-source application that allows practitioners to probe, visualize, and analyze ML systems, with minimal coding. The What-If Tool lets practitioners test performance in hypothetical situations, analyze the importance of different data features, and visualize model behavior across multiple models and subsets of input data. It also lets practitioners measure systems according to multiple ML fairness metrics. We describe the design of the tool, and report on real-life usage at different organizations.
Tasks
Published 2019-07-09
URL https://arxiv.org/abs/1907.04135v2
PDF https://arxiv.org/pdf/1907.04135v2.pdf
PWC https://paperswithcode.com/paper/the-what-if-tool-interactive-probing-of
Repo
Framework

Deep Trajectory for Recognition of Human Behaviours

Title Deep Trajectory for Recognition of Human Behaviours
Authors Tauseef Ali, Eissa Jaber Alreshidi
Abstract Identifying human actions in complex scenes is widely considered as a challenging research problem due to the unpredictable behaviors and variation of appearances and postures. For extracting variations in motion and postures, trajectories provide meaningful way. However, simple trajectories are normally represented by vector of spatial coordinates. In order to identify human actions, we must exploit structural relationship between different trajectories. In this paper, we propose a method that divides the video into N number of segments and then for each segment we extract trajectories. We then compute trajectory descriptor for each segment which capture the structural relationship among different trajectories in the video segment. For trajectory descriptor, we project all extracted trajectories on the canvas. This will result in texture image which can store the relative motion and structural relationship among the trajectories. We then train Convolution Neural Network (CNN) to capture and learn the representation from dense trajectories. . Experimental results shows that our proposed method out performs state of the art methods by 90.01% on benchmark data set.
Tasks
Published 2019-05-24
URL https://arxiv.org/abs/1905.10357v1
PDF https://arxiv.org/pdf/1905.10357v1.pdf
PWC https://paperswithcode.com/paper/deep-trajectory-for-recognition-of-human
Repo
Framework

On Approximate Nonlinear Gaussian Message Passing On Factor Graphs

Title On Approximate Nonlinear Gaussian Message Passing On Factor Graphs
Authors Eike Petersen, Christian Hoffmann, Philipp Rostalski
Abstract Factor graphs have recently gained increasing attention as a unified framework for representing and constructing algorithms for signal processing, estimation, and control. One capability that does not seem to be well explored within the factor graph tool kit is the ability to handle deterministic nonlinear transformations, such as those occurring in nonlinear filtering and smoothing problems, using tabulated message passing rules. In this contribution, we provide general forward (filtering) and backward (smoothing) approximate Gaussian message passing rules for deterministic nonlinear transformation nodes in arbitrary factor graphs fulfilling a Markov property, based on numerical quadrature procedures for the forward pass and a Rauch-Tung-Striebel-type approximation of the backward pass. These message passing rules can be employed for deriving many algorithms for solving nonlinear problems using factor graphs, as is illustrated by the proposition of a nonlinear modified Bryson-Frazier (MBF) smoother based on the presented message passing rules.
Tasks
Published 2019-03-21
URL http://arxiv.org/abs/1903.09136v1
PDF http://arxiv.org/pdf/1903.09136v1.pdf
PWC https://paperswithcode.com/paper/on-approximate-nonlinear-gaussian-message
Repo
Framework

A New Corpus for Low-Resourced Sindhi Language with Word Embeddings

Title A New Corpus for Low-Resourced Sindhi Language with Word Embeddings
Authors Wazir Ali, Jay Kumar, Junyu Lu, Zenglin Xu
Abstract Representing words and phrases into dense vectors of real numbers which encode semantic and syntactic properties is a vital constituent in natural language processing (NLP). The success of neural network (NN) models in NLP largely rely on such dense word representations learned on the large unlabeled corpus. Sindhi is one of the rich morphological language, spoken by large population in Pakistan and India lacks corpora which plays an essential role of a test-bed for generating word embeddings and developing language independent NLP systems. In this paper, a large corpus of more than 61 million words is developed for low-resourced Sindhi language for training neural word embeddings. The corpus is acquired from multiple web-resources using web-scrappy. Due to the unavailability of open source preprocessing tools for Sindhi, the prepossessing of such large corpus becomes a challenging problem specially cleaning of noisy data extracted from web resources. Therefore, a preprocessing pipeline is employed for the filtration of noisy text. Afterwards, the cleaned vocabulary is utilized for training Sindhi word embeddings with state-of-the-art GloVe, Skip-Gram (SG), and Continuous Bag of Words (CBoW) word2vec algorithms. The intrinsic evaluation approach of cosine similarity matrix and WordSim-353 are employed for the evaluation of generated Sindhi word embeddings. Moreover, we compare the proposed word embeddings with recently revealed Sindhi fastText (SdfastText) word representations. Our intrinsic evaluation results demonstrate the high quality of our generated Sindhi word embeddings using SG, CBoW, and GloVe as compare to SdfastText word representations.
Tasks Word Embeddings
Published 2019-11-28
URL https://arxiv.org/abs/1911.12579v2
PDF https://arxiv.org/pdf/1911.12579v2.pdf
PWC https://paperswithcode.com/paper/a-new-corpus-for-low-resourced-sindhi
Repo
Framework

INSET: Sentence Infilling with Inter-sentential Generative Pre-training

Title INSET: Sentence Infilling with Inter-sentential Generative Pre-training
Authors Yichen Huang, Yizhe Zhang, Oussama Elachqar, Yu Cheng
Abstract Missing sentence generation (or sentence infilling) fosters a wide range of applications in natural language generation, such as document auto-completion and meeting note expansion. Such a task asks the model to generate intermediate missing sentence that can semantically and syntactically bridge the surrounding context. Solving the sentence infilling task requires techniques in NLP ranging from natural language understanding, discourse-level planning and natural language generation. In this paper, we present a framework to decouple this challenge and address these three aspects respectively, leveraging the power of existing large-scale pre-trained models such as BERT and GPT-2. Our empirical results demonstrate the effectiveness of our proposed model in learning a sentence representation for generation, and further generating a missing sentence that bridges the context.
Tasks Text Generation
Published 2019-11-10
URL https://arxiv.org/abs/1911.03892v1
PDF https://arxiv.org/pdf/1911.03892v1.pdf
PWC https://paperswithcode.com/paper/inset-sentence-infilling-with-inter
Repo
Framework

Bag of Negatives for Siamese Architectures

Title Bag of Negatives for Siamese Architectures
Authors Bojana Gajic, Ariel Amato, Ramon Baldrich, Carlo Gatta
Abstract Training a Siamese architecture for re-identification with a large number of identities is a challenging task due to the difficulty of finding relevant negative samples efficiently. In this work we present Bag of Negatives (BoN), a method for accelerated and improved training of Siamese networks that scales well on datasets with a very large number of identities. BoN is an efficient and loss-independent method, able to select a bag of high quality negatives, based on a novel online hashing strategy.
Tasks
Published 2019-08-06
URL https://arxiv.org/abs/1908.02391v1
PDF https://arxiv.org/pdf/1908.02391v1.pdf
PWC https://paperswithcode.com/paper/bag-of-negatives-for-siamese-architectures
Repo
Framework

An empirical study of neural networks for trend detection in time series

Title An empirical study of neural networks for trend detection in time series
Authors Alexandre Miot, Gilles Drigout
Abstract Detecting structure in noisy time series is a difficult task. One intuitive feature is the notion of trend. From theoretical hints and using simulated time series, we empirically investigate the efficiency of standard recurrent neural networks (RNNs) to detect trends. We show the overall superiority and versatility of certain standard RNNs structures over various other estimators. These RNNs could be used as basic blocks to build more complex time series trend estimators.
Tasks Time Series
Published 2019-12-09
URL https://arxiv.org/abs/1912.04009v2
PDF https://arxiv.org/pdf/1912.04009v2.pdf
PWC https://paperswithcode.com/paper/an-empirical-study-of-neural-networks-for
Repo
Framework
Title Reporting on Decision-Making Algorithms and some Related Ethical Questions
Authors Benoît Otjacques
Abstract Companies report on their financial performance for decades. More recently they have also started to report on their environmental impact and their social responsibility. The latest trend is now to deliver one single integrated report where all stakeholders of the company can easily connect all facets of the business with their impact considered in a broad sense. The main purpose of this integrated approach is to avoid delivering data related to disconnected silos, which consequently makes it very difficult to globally assess the overall performance of an entity or a business line. In this paper, we focus on how companies report on risks and ethical issues related to the increasing use of Artificial Intelligence (AI). We explain some of these risks and potential issues. Next, we identify some recent initiatives by various stakeholders to define a global ethical framework for AI. Finally, we illustrate with four cases that companies are very shy to report on these facets of AI.
Tasks Decision Making
Published 2019-11-04
URL https://arxiv.org/abs/1911.05731v1
PDF https://arxiv.org/pdf/1911.05731v1.pdf
PWC https://paperswithcode.com/paper/reporting-on-decision-making-algorithms-and
Repo
Framework
Title AiAds: Automated and Intelligent Advertising System for Sponsored Search
Authors Xiao Yang, Daren Sun, Ruiwei Zhu, Tao Deng, Zhi Guo, Jiao Ding, Shouke Qin, Zongyao Ding, Yanfeng Zhu
Abstract Sponsored search has more than 20 years of history, and it has been proven to be a successful business model for online advertising. Based on the pay-per-click pricing model and the keyword targeting technology, the sponsored system runs online auctions to determine the allocations and prices of search advertisements. In the traditional setting, advertisers should manually create lots of ad creatives and bid on some relevant keywords to target their audience. Due to the huge amount of search traffic and a wide variety of ad creations, the limits of manual optimizations from advertisers become the main bottleneck for improving the efficiency of this market. Moreover, as many emerging advertising forms and supplies are growing, it’s crucial for sponsored search platform to pay more attention to the ROI metrics of ads for getting the marketing budgets of advertisers. In this paper, we present the AiAds system developed at Baidu, which use machine learning techniques to build an automated and intelligent advertising system. By designing and implementing the automated bidding strategy, the intelligent targeting and the intelligent creation models, the AiAds system can transform the manual optimizations into multiple automated tasks and optimize these tasks in advanced methods. AiAds is a brand-new architecture of sponsored search system which changes the bidding language and allocation mechanism, breaks the limit of keyword targeting with end-to-end ad retrieval framework and provides global optimization of ad creation. This system can increase the advertiser’s campaign performance, the user experience and the revenue of the advertising platform simultaneously and significantly. We present the overall architecture and modeling techniques for each module of the system and share our lessons learned in solving several key challenges.
Tasks
Published 2019-07-28
URL https://arxiv.org/abs/1907.12118v1
PDF https://arxiv.org/pdf/1907.12118v1.pdf
PWC https://paperswithcode.com/paper/aiads-automated-and-intelligent-advertising
Repo
Framework

Investigating Flight Envelope Variation Predictability of Impaired Aircraft using Least-Squares Regression Analysis

Title Investigating Flight Envelope Variation Predictability of Impaired Aircraft using Least-Squares Regression Analysis
Authors Ramin Norouzi, Amirreza Kosari, Mohammad Hossein Sabour
Abstract Aircraft failures alter the aircraft dynamics and cause maneuvering flight envelope to change. Such envelope variations are nonlinear and generally unpredictable by the pilot as they are governed by the aircraft’s complex dynamics. Hence, in order to prevent in-flight Loss of Control it is crucial to practically predict the impaired aircraft’s flight envelope variation due to any a-priori unknown failure degree. This paper investigates the predictability of the number of trim points within the maneuvering flight envelope and its centroid using both linear and nonlinear least-squares estimation methods. To do so, various polynomial models and nonlinear models based on hyperbolic tangent function are developed and compared which incorporate the influencing factors on the envelope variations as the inputs and estimate the centroid and the number of trim points of the maneuvering flight envelope at any intended failure degree. Results indicate that both the polynomial and hyperbolic tangent function-based models are capable of predicting the impaired fight envelope variation with good precision. Furthermore, it is shown that the regression equation of the best polynomial fit enables direct assessment of the impaired aircraft’s flight envelope contraction and displacement sensitivity to the specific parameters characterizing aircraft failure and flight condition.
Tasks
Published 2019-05-20
URL https://arxiv.org/abs/1905.07875v2
PDF https://arxiv.org/pdf/1905.07875v2.pdf
PWC https://paperswithcode.com/paper/investigating-flight-envelope-variation
Repo
Framework

Sentence-Level Content Planning and Style Specification for Neural Text Generation

Title Sentence-Level Content Planning and Style Specification for Neural Text Generation
Authors Xinyu Hua, Lu Wang
Abstract Building effective text generation systems requires three critical components: content selection, text planning, and surface realization, and traditionally they are tackled as separate problems. Recent all-in-one style neural generation models have made impressive progress, yet they often produce outputs that are incoherent and unfaithful to the input. To address these issues, we present an end-to-end trained two-step generation model, where a sentence-level content planner first decides on the keyphrases to cover as well as a desired language style, followed by a surface realization decoder that generates relevant and coherent text. For experiments, we consider three tasks from domains with diverse topics and varying language styles: persuasive argument construction from Reddit, paragraph generation for normal and simple versions of Wikipedia, and abstract generation for scientific articles. Automatic evaluation shows that our system can significantly outperform competitive comparisons. Human judges further rate our system generated text as more fluent and correct, compared to the generations by its variants that do not consider language style.
Tasks Text Generation
Published 2019-09-02
URL https://arxiv.org/abs/1909.00734v1
PDF https://arxiv.org/pdf/1909.00734v1.pdf
PWC https://paperswithcode.com/paper/sentence-level-content-planning-and-style
Repo
Framework

Data adaptation in HANDY economy-ideology model

Title Data adaptation in HANDY economy-ideology model
Authors Marcin Sendera
Abstract The concept of mathematical modeling is widespread across almost all of the fields of contemporary science and engineering. Because of the existing necessity of predictions the behavior of natural phenomena, the researchers develop more and more complex models. However, despite their ability to better forecasting, the problem of an appropriate fitting ground truth data to those, high-dimensional and nonlinear models seems to be inevitable. In order to deal with this demanding problem the entire discipline of data assimilation has been developed. Basing on the Human and Nature Dynamics (HANDY) model, we have presented a detailed and comprehensive comparison of Approximate Bayesian Computation (classic data assimilation method) and a novelty approach of Supermodeling. Furthermore, with the usage of Sensitivity Analysis, we have proposed the methodology to reduce the number of coupling coefficients between submodels and as a consequence to increase the speed of the Supermodel converging. In addition, we have demonstrated that usage of Approximate Bayesian Computation method with the knowledge about parameters’ sensitivities could result with satisfactory estimation of the initial parameters. However, we have also presented the mentioned methodology as unable to achieve similar predictions to Approximate Bayesian Computation. Finally, we have proved that Supermodeling with synchronization via the most sensitive variable could effect with the better forecasting for chaotic as well as more stable systems than the Approximate Bayesian Computation. What is more, we have proposed the adequate methodologies.
Tasks
Published 2019-04-08
URL http://arxiv.org/abs/1904.04309v1
PDF http://arxiv.org/pdf/1904.04309v1.pdf
PWC https://paperswithcode.com/paper/data-adaptation-in-handy-economy-ideology
Repo
Framework

Skew-Fit: State-Covering Self-Supervised Reinforcement Learning

Title Skew-Fit: State-Covering Self-Supervised Reinforcement Learning
Authors Vitchyr H. Pong, Murtaza Dalal, Steven Lin, Ashvin Nair, Shikhar Bahl, Sergey Levine
Abstract Autonomous agents that must exhibit flexible and broad capabilities will need to be equipped with large repertoires of skills. Defining each skill with a manually-designed reward function limits this repertoire and imposes a manual engineering burden. Self-supervised agents that set their own goals can automate this process, but designing appropriate goal setting objectives can be difficult, and often involves heuristic design decisions. In this paper, we propose a formal exploration objective for goal-reaching policies that maximizes state coverage. We show that this objective is equivalent to maximizing goal reaching performance together with the entropy of the goal distribution, where goals correspond to full state observations. To instantiate this principle, we present an algorithm called Skew-Fit for learning a maximum-entropy goal distributions. We prove that, under regularity conditions, Skew-Fit converges to a uniform distribution over the set of valid states, even when we do not know this set beforehand. Our experiments show that combining Skew-Fit for learning goal distributions with existing goal-reaching methods outperforms a variety of prior methods on open-sourced visual goal-reaching tasks. Moreover, we demonstrate that \METHOD enables a real-world robot to learn to open a door, entirely from scratch, from pixels, and without any manually-designed reward function.
Tasks
Published 2019-03-08
URL https://arxiv.org/abs/1903.03698v3
PDF https://arxiv.org/pdf/1903.03698v3.pdf
PWC https://paperswithcode.com/paper/skew-fit-state-covering-self-supervised
Repo
Framework
comments powered by Disqus