January 25, 2020

3394 words 16 mins read

Paper Group NAWR 4

Paper Group NAWR 4

Automatic Inference of Sound Correspondence Patterns across Multiple Languages. Cascade of Encoder-Decoder CNNs with Learned Coordinates Regressor for Robust Facial Landmarks Detection. A unified theory for the origin of grid cells through the lens of pattern formation. Coreference Resolution with Entity Equalization. ODE2VAE: Deep generative secon …

Automatic Inference of Sound Correspondence Patterns across Multiple Languages

Title Automatic Inference of Sound Correspondence Patterns across Multiple Languages
Authors Johann-Mattis List
Abstract Sound correspondence patterns play a crucial role for linguistic reconstruction. Linguists use them to prove language relationship, to reconstruct proto-forms, and for classical phylogenetic reconstruction based on shared innovations. Cognate words that fail to conform with expected patterns can further point to various kinds of exceptions in sound change, such as analogy or assimilation of frequent words. Here I present an automatic method for the inference of sound correspondence patterns across multiple languages based on a network approach. The core idea is to represent all columns in aligned cognate sets as nodes in a network with edges representing the degree of compatibility between the nodes. The task of inferring all compatible correspondence sets can then be handled as the well-known minimum clique cover problem in graph theory, which essentially seeks to split the graph into the smallest number of cliques in which each node is represented by exactly one clique. The resulting partitions represent all correspondence patterns that can be inferred for a given data set. By excluding those patterns that occur in only a few cognate sets, the core of regularly recurring sound correspondences can be inferred. Based on this idea, the article presents a method for automatic correspondence pattern recognition, which is implemented as part of a Python library which supplements the article. To illustrate the usefulness of the method, I present how the inferred patterns can be used to predict words that have not been observed before.
Tasks
Published 2019-03-01
URL https://www.aclweb.org/anthology/J19-1004/
PDF https://www.aclweb.org/anthology/J19-1004
PWC https://paperswithcode.com/paper/automatic-inference-of-sound-correspondence
Repo https://github.com/lingpy/correspondence-pattern-paper
Framework none

Cascade of Encoder-Decoder CNNs with Learned Coordinates Regressor for Robust Facial Landmarks Detection

Title Cascade of Encoder-Decoder CNNs with Learned Coordinates Regressor for Robust Facial Landmarks Detection
Authors Roberto Valle, Jose M. Buenaposada, Luis Baumela
Abstract Convolutional Neural Nets (CNNs) have become the reference technology for many computer vision problems. Although CNNs for facial landmark detection are very robust, they still lack accuracy when processing images acquired in unrestricted conditions. In this paper we investigate the use of a cascade of Neural Net regressors to increase the accuracy of the estimated facial landmarks. To this end we append two encoder-decoder CNNs with the same architecture. The first net produces a set of heatmaps with a rough estimation of landmark locations. The second, trained with synthetically generated occlusions, refines the location of ambiguous and occluded landmarks. Finally, a densely connected layer with shared weights among all heatmaps, accurately regresses the landmark coordinates. The proposed approach achieves state-of-the-art results in 300W, COFW and WFLW that are widely considered the most challenging public data sets.
Tasks Face Alignment, Facial Landmark Detection
Published 2019-10-15
URL https://doi.org/10.1016/j.patrec.2019.10.012
PDF https://jmbuena.github.io/publications/PRL_2019_web.pdf
PWC https://paperswithcode.com/paper/cascade-of-encoder-decoder-cnns-with-learned
Repo https://github.com/bobetocalo/bobetocalo_prl19
Framework none

A unified theory for the origin of grid cells through the lens of pattern formation

Title A unified theory for the origin of grid cells through the lens of pattern formation
Authors Ben Sorscher, Gabriel Mel, Surya Ganguli, Samuel Ocko
Abstract Grid cells in the brain fire in strikingly regular hexagonal patterns across space. There are currently two seemingly unrelated frameworks for understanding these patterns. Mechanistic models account for hexagonal firing fields as the result of pattern-forming dynamics in a recurrent neural network with hand-tuned center-surround connectivity. Normative models specify a neural architecture, a learning rule, and a navigational task, and observe that grid-like firing fields emerge due to the constraints of solving this task. Here we provide an analytic theory that unifies the two perspectives by casting the learning dynamics of neural networks trained on navigational tasks as a pattern forming dynamical system. This theory provides insight into the optimal solutions of diverse formulations of the normative task, and shows that symmetries in the representation of space correctly predict the structure of learned firing fields in trained neural networks. Further, our theory proves that a nonnegativity constraint on firing rates induces a symmetry-breaking mechanism which favors hexagonal firing fields. We extend this theory to the case of learning multiple grid maps and demonstrate that optimal solutions consist of a hierarchy of maps with increasing length scales. These results unify previous accounts of grid cell firing and provide a novel framework for predicting the learned representations of recurrent neural networks.
Tasks
Published 2019-12-01
URL http://papers.nips.cc/paper/9191-a-unified-theory-for-the-origin-of-grid-cells-through-the-lens-of-pattern-formation
PDF http://papers.nips.cc/paper/9191-a-unified-theory-for-the-origin-of-grid-cells-through-the-lens-of-pattern-formation.pdf
PWC https://paperswithcode.com/paper/a-unified-theory-for-the-origin-of-grid-cells
Repo https://github.com/ganguli-lab/grid-pattern-formation
Framework tf

Coreference Resolution with Entity Equalization

Title Coreference Resolution with Entity Equalization
Authors Ben Kantor, Amir Globerson
Abstract A key challenge in coreference resolution is to capture properties of entity clusters, and use those in the resolution process. Here we provide a simple and effective approach for achieving this, via an {``}Entity Equalization{''} mechanism. The Equalization approach represents each mention in a cluster via an approximation of the sum of all mentions in the cluster. We show how this can be done in a fully differentiable end-to-end manner, thus enabling high-order inferences in the resolution process. Our approach, which also employs BERT embeddings, results in new state-of-the-art results on the CoNLL-2012 coreference resolution task, improving average F1 by 3.6{%}. |
Tasks Coreference Resolution
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1066/
PDF https://www.aclweb.org/anthology/P19-1066
PWC https://paperswithcode.com/paper/coreference-resolution-with-entity
Repo https://github.com/kkjawz/coref-ee
Framework tf

ODE2VAE: Deep generative second order ODEs with Bayesian neural networks

Title ODE2VAE: Deep generative second order ODEs with Bayesian neural networks
Authors Cagatay Yildiz, Markus Heinonen, Harri Lahdesmaki
Abstract We present Ordinary Differential Equation Variational Auto-Encoder (ODE2VAE), a latent second order ODE model for high-dimensional sequential data. Leveraging the advances in deep generative models, ODE2VAE can simultaneously learn the embedding of high dimensional trajectories and infer arbitrarily complex continuous-time latent dynamics. Our model explicitly decomposes the latent space into momentum and position components and solves a second order ODE system, which is in contrast to recurrent neural network (RNN) based time series models and recently proposed black-box ODE techniques. In order to account for uncertainty, we propose probabilistic latent ODE dynamics parameterized by deep Bayesian neural networks. We demonstrate our approach on motion capture, image rotation, and bouncing balls datasets. We achieve state-of-the-art performance in long term motion prediction and imputation tasks.
Tasks Imputation, Motion Capture, motion prediction, Time Series
Published 2019-12-01
URL http://papers.nips.cc/paper/9497-ode2vae-deep-generative-second-order-odes-with-bayesian-neural-networks
PDF http://papers.nips.cc/paper/9497-ode2vae-deep-generative-second-order-odes-with-bayesian-neural-networks.pdf
PWC https://paperswithcode.com/paper/ode2vae-deep-generative-second-order-odes-1
Repo https://github.com/cagatayyildiz/ODE2VAE
Framework tf

As-projective-as-possible bias correction for illumination estimation algorithms

Title As-projective-as-possible bias correction for illumination estimation algorithms
Authors Mahmoud Afifi, Abhijith Punnappurath, Graham Finlayson, Michael S. Brown
Abstract Illumination estimation is the key routine in a camera’s onboard auto-white-balance (AWB) function. Illumination estimation algorithms estimate the color of the scene’s illumination from an image in the form of an R,G,B vector in the sensor’s raw-RGB color space. While learning-based methods have demonstrated impressive performance for illumination estimation, cameras still rely on simple statistical-based algorithms that are less accurate but capable of executing quickly on the camera’s hardware. An effective strategy to improve the accuracy of these fast statistical-based algorithms is to apply a post-estimate bias correction function to transform the estimated R,G,B vector such that it lies closer to the correct solution. Recent work by Finlayson, Interface Focus, 2018 showed that a bias correction function can be formulated as a projective transform because the magnitude of the R,G,B illumination vector does not matter to the AWB procedure. This paper builds on this finding and shows that further improvements can be obtained by using an as-projective-as-possible (APAP) projective transform that locally adapts the projective transform to the input R,G,B vector. We demonstrate the effectiveness of the proposed APAP bias correction on several well-known statistical illumination estimation methods. We also describe a fast lookup method that allows the APAP transform to be performed with only a few lookup operations.
Tasks
Published 2019-01-01
URL https://www.osapublishing.org/josaa/abstract.cfm?uri=josaa-36-1-71
PDF http://www.cse.yorku.ca/~mbrown/pdf/josa_2019_afifi.pdf
PWC https://paperswithcode.com/paper/as-projective-as-possible-bias-correction-for
Repo https://github.com/mahmoudnafifi/APAP-bias-correction-for-illumination-estimation-methods
Framework none

GraphDTA: prediction of drug–target binding affinity using graph convolutional networks

Title GraphDTA: prediction of drug–target binding affinity using graph convolutional networks
Authors Thin Nguyen, Hang Le, Svetha Venkatesh
Abstract While the development of new drugs is costly, time consuming, and often accompanied with safety issues, drug repurposing, where old drugs with established safety are used for medical conditions other than originally developed, is an attractive alternative. Then, how the old drugs work on new targets becomes a crucial part of drug repurposing and gains much of interest. Several statistical and machine learning models have been proposed to estimate drug–target binding affinity and deep learning approaches have been shown to be among state-of-the-art methods. However, drugs and targets in these models were commonly represented in 1D strings, regardless the fact that molecules are by nature formed by the chemical bonding of atoms. In this work, we propose GraphDTA to capture the structural information of drugs, possibly enhancing the predictive power of the affinity. In particular, unlike competing methods, drugs are represented as graphs and graph convolutional networks are used to learn drug–target binding affinity. We trial our method on two benchmark drug–target binding affinity datasets and compare the performance with state-of-the-art models in the field. The results show that our proposed method can not only predict the affinity better than non-deep learning models, but also outperform competing deep learning approaches. This demonstrates the practical advantages of graph-based representation for molecules in providing accurate prediction of drug–target binding affinity. The application may also include any recommendation systems where either or both of the user- and product-like sides can be represented in graphs.
Tasks Recommendation Systems
Published 2019-07-02
URL https://www.biorxiv.org/content/10.1101/684662v3
PDF https://www.biorxiv.org/content/biorxiv/early/2019/07/02/684662.full.pdf
PWC https://paperswithcode.com/paper/graphdta-prediction-of-drugtarget-binding
Repo https://github.com/thinng/GraphDTA
Framework pytorch

Event Detection with Trigger-Aware Lattice Neural Network

Title Event Detection with Trigger-Aware Lattice Neural Network
Authors Ning Ding, Ziran Li, Zhiyuan Liu, Haitao Zheng, Zibo Lin
Abstract Event detection (ED) aims to locate trigger words in raw text and then classify them into correct event types. In this task, neural net- work based models became mainstream in re- cent years. However, two problems arise when it comes to languages without natural delim- iters, such as Chinese. First, word-based mod- els severely suffer from the problem of word- trigger mismatch, limiting the performance of the methods. In addition, even if trigger words could be accurately located, the ambi- guity of polysemy of triggers could still af- fect the trigger classification stage. To ad- dress the two issues simultaneously, we pro- pose the Trigger-aware Lattice Neural Net- work (TLNN). (1) The framework dynami- cally incorporates word and character informa- tion so that the trigger-word mismatch issue can be avoided. (2) Moreover, for polysemous characters and words, we model all senses of them with the help of an external linguistic knowledge base, so as to alleviate the prob- lem of ambiguous triggers. Experiments on two benchmark datasets show that our model could effectively tackle the two issues and outperforms previous state-of-the-art methods significantly, giving the best results. The source code of this paper can be obtained from https://github.com/thunlp/TLNN.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1033/
PDF https://www.aclweb.org/anthology/D19-1033
PWC https://paperswithcode.com/paper/event-detection-with-trigger-aware-lattice
Repo https://github.com/thunlp/TLNN
Framework pytorch

Actions Speak Louder than Goals:Valuing Player Actions in Soccer

Title Actions Speak Louder than Goals:Valuing Player Actions in Soccer
Authors Tom Decroos, Lotte Bransen, Jan Van Haaren, Jesse Davis
Abstract Assessing the impact of the individual actions performed by soccerplayers during games is a crucial aspect of the player recruitmentprocess. Unfortunately, most traditional metrics fall short in ad-dressing this task as they either focus on rare actions like shotsand goals alone or fail to account for the context in which theactions occurred. This paper introduces (1) a new language for de-scribing individual player actions on the pitch and (2) a frameworkfor valuing any type of player action based on its impact on thegame outcome while accounting for the context in which the actionhappened. By aggregating soccer players’ action values, their totaloffensive and defensive contributions to their team can be quan-tified. We show how our approach considers relevant contextualinformation that traditional player evaluation metrics ignore andpresent a number of use cases related to scouting and playing stylecharacterization in the 2016/2017 and 2017/2018 seasons in Europe’stop competitions.
Tasks Football Action Valuation
Published 2019-07-10
URL https://www.researchgate.net/publication/323302738_Actions_Speak_Louder_Than_Goals_Valuing_Player_Actions_in_Soccer
PDF https://arxiv.org/pdf/1802.07127.pdf
PWC https://paperswithcode.com/paper/actions-speak-louder-than-goalsvaluing-player
Repo https://github.com/ML-KULeuven/socceraction
Framework none

Unsupervised Discovery of Temporal Structure in Noisy Data with Dynamical Components Analysis

Title Unsupervised Discovery of Temporal Structure in Noisy Data with Dynamical Components Analysis
Authors David Clark, Jesse Livezey, Kristofer Bouchard
Abstract Linear dimensionality reduction methods are commonly used to extract low-dimensional structure from high-dimensional data. However, popular methods disregard temporal structure, rendering them prone to extracting noise rather than meaningful dynamics when applied to time series data. At the same time, many successful unsupervised learning methods for temporal, sequential and spatial data extract features which are predictive of their surrounding context. Combining these approaches, we introduce Dynamical Components Analysis (DCA), a linear dimensionality reduction method which discovers a subspace of high-dimensional time series data with maximal predictive information, defined as the mutual information between the past and future. We test DCA on synthetic examples and demonstrate its superior ability to extract dynamical structure compared to commonly used linear methods. We also apply DCA to several real-world datasets, showing that the dimensions extracted by DCA are more useful than those extracted by other methods for predicting future states and decoding auxiliary variables. Overall, DCA robustly extracts dynamical structure in noisy, high-dimensional data while retaining the computational efficiency and geometric interpretability of linear dimensionality reduction methods.
Tasks Dimensionality Reduction, Time Series
Published 2019-12-01
URL http://papers.nips.cc/paper/9574-unsupervised-discovery-of-temporal-structure-in-noisy-data-with-dynamical-components-analysis
PDF http://papers.nips.cc/paper/9574-unsupervised-discovery-of-temporal-structure-in-noisy-data-with-dynamical-components-analysis.pdf
PWC https://paperswithcode.com/paper/unsupervised-discovery-of-temporal-structure-1
Repo https://github.com/BouchardLab/DynamicalComponentsAnalysis
Framework none

Joint Acne Image Grading and Counting via Label Distribution Learning

Title Joint Acne Image Grading and Counting via Label Distribution Learning
Authors Xiaoping Wu, Ni Wen, Jie Liang, Yu-Kun Lai, Dongyu She, Ming-Ming Cheng, Jufeng Yang
Abstract Accurate grading of skin disease severity plays a crucial role in precise treatment for patients. Acne vulgaris, the most common skin disease in adolescence, can be graded by evidence-based lesion counting as well as experience-based global estimation in the medical field. However, due to the appearance similarity of acne with close severity, it is challenging to count and grade acne accurately. In this paper, we address the problem of acne image analysis via Label Distribution Learning (LDL) considering the ambiguous information among acne severity. Based on the professional grading criterion, we generate two acne label distributions considering the relationship between the similar number of lesions and severity of acne, respectively. We also propose a unified framework for joint acne image grading and counting, which is optimized by the multi-task learning loss. In addition, we further build the ACNE04 dataset with annotations of acne severity and lesion number of each image for evaluation. Experiments demonstrate that our proposed framework performs favorably against state-of-the-art methods. We make the code and dataset publicly available at https://github.com/xpwu95/ldl.
Tasks Multi-Task Learning
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Wu_Joint_Acne_Image_Grading_and_Counting_via_Label_Distribution_Learning_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Wu_Joint_Acne_Image_Grading_and_Counting_via_Label_Distribution_Learning_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/joint-acne-image-grading-and-counting-via
Repo https://github.com/xpwu95/ldl
Framework pytorch

CNN^{2}: Viewpoint Generalization via a Binocular Vision

Title CNN^{2}: Viewpoint Generalization via a Binocular Vision
Authors Wei-Da Chen, Shan-Hung (Brandon) Wu
Abstract The Convolutional Neural Networks (CNNs) have laid the foundation for many techniques in various applications. Despite achieving remarkable performance in some tasks, the 3D viewpoint generalizability of CNNs is still far behind humans visual capabilities. Although recent efforts, such as the Capsule Networks, have been made to address this issue, these new models are either hard to train and/or incompatible with existing CNN-based techniques specialized for different applications. Observing that humans use binocular vision to understand the world, we study in this paper whether the 3D viewpoint generalizability of CNNs can be achieved via a binocular vision. We propose CNN^{2}, a CNN that takes two images as input, which resembles the process of an object being viewed from the left eye and the right eye. CNN^{2} uses novel augmentation, pooling, and convolutional layers to learn a sense of three-dimensionality in a recursive manner. Empirical evaluation shows that CNN^{2} has improved viewpoint generalizability compared to vanilla CNNs. Furthermore, CNN^{2} is easy to implement and train, and is compatible with existing CNN-based specialized techniques for different applications.
Tasks
Published 2019-12-01
URL http://papers.nips.cc/paper/8473-cnn2-viewpoint-generalization-via-a-binocular-vision
PDF http://papers.nips.cc/paper/8473-cnn2-viewpoint-generalization-via-a-binocular-vision.pdf
PWC https://paperswithcode.com/paper/cnn2-viewpoint-generalization-via-a-binocular
Repo https://github.com/wdchenxyz/CNN2
Framework tf

Using Natural Language Relations between Answer Choices for Machine Comprehension

Title Using Natural Language Relations between Answer Choices for Machine Comprehension
Authors Rajkumar Pujari, Dan Goldwasser
Abstract While evaluating an answer choice for Reading Comprehension task, other answer choices available for the question and the answers of related questions about the same paragraph often provide valuable information. In this paper, we propose a method to leverage the natural language relations between the answer choices, such as entailment and contradiction, to improve the performance of machine comprehension. We use a stand-alone question answering (QA) system to perform QA task and a Natural Language Inference (NLI) system to identify the relations between the choice pairs. Then we perform inference using an Integer Linear Programming (ILP)-based relational framework to re-evaluate the decisions made by the standalone QA system in light of the relations identified by the NLI system. We also propose a multitask learning model that learns both the tasks jointly.
Tasks Natural Language Inference, Question Answering, Reading Comprehension
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-1403/
PDF https://www.aclweb.org/anthology/N19-1403
PWC https://paperswithcode.com/paper/using-natural-language-relations-between
Repo https://github.com/pujari-rajkumar/naacl2019
Framework none

Scalable Collapsed Inference for High-Dimensional Topic Models

Title Scalable Collapsed Inference for High-Dimensional Topic Models
Authors Rashidul Islam, James Foulds
Abstract The bigger the corpus, the more topics it can potentially support. To truly make full use of massive text corpora, a topic model inference algorithm must therefore scale efficiently in 1) documents and 2) topics, while 3) achieving accurate inference. Previous methods have achieved two out of three of these criteria simultaneously, but never all three at once. In this paper, we develop an online inference algorithm for topic models which leverages stochasticity to scale well in the number of documents, sparsity to scale well in the number of topics, and which operates in the collapsed representation of the topic model for improved accuracy and run-time performance. We use a Monte Carlo inner loop in the online setting to approximate the collapsed variational Bayes updates in a sparse and efficient way, which we accomplish via the MetropolisHastings Walker method. We showcase our algorithm on LDA and the recently proposed mixed membership skip-gram topic model. Our method requires only amortized $O(k_{d})$ computation per word token instead of $O(K)$ operations, where the number of topics occurring for a particular document $k_{d}\ll$ the total number of topics in the corpus $K$, to converge to a high-quality solution.
Tasks Topic Models
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-1291/
PDF https://www.aclweb.org/anthology/N19-1291
PWC https://paperswithcode.com/paper/scalable-collapsed-inference-for-high
Repo https://github.com/dr97531/SparseSCVB0
Framework none

Multi-Criteria Dimensionality Reduction with Applications to Fairness

Title Multi-Criteria Dimensionality Reduction with Applications to Fairness
Authors Uthaipon Tantipongpipat, Samira Samadi, Mohit Singh, Jamie H. Morgenstern, Santosh Vempala
Abstract Dimensionality reduction is a classical technique widely used for data analysis. One foundational instantiation is Principal Component Analysis (PCA), which minimizes the average reconstruction error. In this paper, we introduce the multi-criteria dimensionality reduction problem where we are given multiple objectives that need to be optimized simultaneously. As an application, our model captures several fairness criteria for dimensionality reduction such as the Fair-PCA problem introduced by Samadi et al. [NeurIPS18] and the Nash Social Welfare (NSW) problem. In the Fair-PCA problem, the input data is divided into k groups, and the goal is to find a single d-dimensional representation for all groups for which the maximum reconstruction error of any one group is minimized. In NSW the goal is to maximize the product of the individual variances of the groups achieved by the common low-dimensinal space. Our main result is an exact polynomial-time algorithm for the two-criteria dimensionality reduction problem when the two criteria are increasing concave functions. As an application of this result, we obtain a polynomial time algorithm for Fair-PCA for k=2 groups, resolving an open problem of Samadi et al.[NeurIPS18], and a polynomial time algorithm for NSW objective for k=2 groups. We also give approximation algorithms for k>2. Our technical contribution in the above results is to prove new low-rank properties of extreme point solutions to semi-definite programs. We conclude with the results of several experiments indicating improved performance and generalized application of our algorithm on real-world datasets.
Tasks Dimensionality Reduction
Published 2019-12-01
URL http://papers.nips.cc/paper/9652-multi-criteria-dimensionality-reduction-with-applications-to-fairness
PDF http://papers.nips.cc/paper/9652-multi-criteria-dimensionality-reduction-with-applications-to-fairness.pdf
PWC https://paperswithcode.com/paper/multi-criteria-dimensionality-reduction-with
Repo https://github.com/SDPforAll/multiCriteriaDimReduction
Framework none
comments powered by Disqus