Paper Group NAWR 4
Automatic Inference of Sound Correspondence Patterns across Multiple Languages. Cascade of Encoder-Decoder CNNs with Learned Coordinates Regressor for Robust Facial Landmarks Detection. A unified theory for the origin of grid cells through the lens of pattern formation. Coreference Resolution with Entity Equalization. ODE2VAE: Deep generative secon …
Automatic Inference of Sound Correspondence Patterns across Multiple Languages
Title | Automatic Inference of Sound Correspondence Patterns across Multiple Languages |
Authors | Johann-Mattis List |
Abstract | Sound correspondence patterns play a crucial role for linguistic reconstruction. Linguists use them to prove language relationship, to reconstruct proto-forms, and for classical phylogenetic reconstruction based on shared innovations. Cognate words that fail to conform with expected patterns can further point to various kinds of exceptions in sound change, such as analogy or assimilation of frequent words. Here I present an automatic method for the inference of sound correspondence patterns across multiple languages based on a network approach. The core idea is to represent all columns in aligned cognate sets as nodes in a network with edges representing the degree of compatibility between the nodes. The task of inferring all compatible correspondence sets can then be handled as the well-known minimum clique cover problem in graph theory, which essentially seeks to split the graph into the smallest number of cliques in which each node is represented by exactly one clique. The resulting partitions represent all correspondence patterns that can be inferred for a given data set. By excluding those patterns that occur in only a few cognate sets, the core of regularly recurring sound correspondences can be inferred. Based on this idea, the article presents a method for automatic correspondence pattern recognition, which is implemented as part of a Python library which supplements the article. To illustrate the usefulness of the method, I present how the inferred patterns can be used to predict words that have not been observed before. |
Tasks | |
Published | 2019-03-01 |
URL | https://www.aclweb.org/anthology/J19-1004/ |
https://www.aclweb.org/anthology/J19-1004 | |
PWC | https://paperswithcode.com/paper/automatic-inference-of-sound-correspondence |
Repo | https://github.com/lingpy/correspondence-pattern-paper |
Framework | none |
Cascade of Encoder-Decoder CNNs with Learned Coordinates Regressor for Robust Facial Landmarks Detection
Title | Cascade of Encoder-Decoder CNNs with Learned Coordinates Regressor for Robust Facial Landmarks Detection |
Authors | Roberto Valle, Jose M. Buenaposada, Luis Baumela |
Abstract | Convolutional Neural Nets (CNNs) have become the reference technology for many computer vision problems. Although CNNs for facial landmark detection are very robust, they still lack accuracy when processing images acquired in unrestricted conditions. In this paper we investigate the use of a cascade of Neural Net regressors to increase the accuracy of the estimated facial landmarks. To this end we append two encoder-decoder CNNs with the same architecture. The first net produces a set of heatmaps with a rough estimation of landmark locations. The second, trained with synthetically generated occlusions, refines the location of ambiguous and occluded landmarks. Finally, a densely connected layer with shared weights among all heatmaps, accurately regresses the landmark coordinates. The proposed approach achieves state-of-the-art results in 300W, COFW and WFLW that are widely considered the most challenging public data sets. |
Tasks | Face Alignment, Facial Landmark Detection |
Published | 2019-10-15 |
URL | https://doi.org/10.1016/j.patrec.2019.10.012 |
https://jmbuena.github.io/publications/PRL_2019_web.pdf | |
PWC | https://paperswithcode.com/paper/cascade-of-encoder-decoder-cnns-with-learned |
Repo | https://github.com/bobetocalo/bobetocalo_prl19 |
Framework | none |
A unified theory for the origin of grid cells through the lens of pattern formation
Title | A unified theory for the origin of grid cells through the lens of pattern formation |
Authors | Ben Sorscher, Gabriel Mel, Surya Ganguli, Samuel Ocko |
Abstract | Grid cells in the brain fire in strikingly regular hexagonal patterns across space. There are currently two seemingly unrelated frameworks for understanding these patterns. Mechanistic models account for hexagonal firing fields as the result of pattern-forming dynamics in a recurrent neural network with hand-tuned center-surround connectivity. Normative models specify a neural architecture, a learning rule, and a navigational task, and observe that grid-like firing fields emerge due to the constraints of solving this task. Here we provide an analytic theory that unifies the two perspectives by casting the learning dynamics of neural networks trained on navigational tasks as a pattern forming dynamical system. This theory provides insight into the optimal solutions of diverse formulations of the normative task, and shows that symmetries in the representation of space correctly predict the structure of learned firing fields in trained neural networks. Further, our theory proves that a nonnegativity constraint on firing rates induces a symmetry-breaking mechanism which favors hexagonal firing fields. We extend this theory to the case of learning multiple grid maps and demonstrate that optimal solutions consist of a hierarchy of maps with increasing length scales. These results unify previous accounts of grid cell firing and provide a novel framework for predicting the learned representations of recurrent neural networks. |
Tasks | |
Published | 2019-12-01 |
URL | http://papers.nips.cc/paper/9191-a-unified-theory-for-the-origin-of-grid-cells-through-the-lens-of-pattern-formation |
http://papers.nips.cc/paper/9191-a-unified-theory-for-the-origin-of-grid-cells-through-the-lens-of-pattern-formation.pdf | |
PWC | https://paperswithcode.com/paper/a-unified-theory-for-the-origin-of-grid-cells |
Repo | https://github.com/ganguli-lab/grid-pattern-formation |
Framework | tf |
Coreference Resolution with Entity Equalization
Title | Coreference Resolution with Entity Equalization |
Authors | Ben Kantor, Amir Globerson |
Abstract | A key challenge in coreference resolution is to capture properties of entity clusters, and use those in the resolution process. Here we provide a simple and effective approach for achieving this, via an {``}Entity Equalization{''} mechanism. The Equalization approach represents each mention in a cluster via an approximation of the sum of all mentions in the cluster. We show how this can be done in a fully differentiable end-to-end manner, thus enabling high-order inferences in the resolution process. Our approach, which also employs BERT embeddings, results in new state-of-the-art results on the CoNLL-2012 coreference resolution task, improving average F1 by 3.6{%}. | |
Tasks | Coreference Resolution |
Published | 2019-07-01 |
URL | https://www.aclweb.org/anthology/P19-1066/ |
https://www.aclweb.org/anthology/P19-1066 | |
PWC | https://paperswithcode.com/paper/coreference-resolution-with-entity |
Repo | https://github.com/kkjawz/coref-ee |
Framework | tf |
ODE2VAE: Deep generative second order ODEs with Bayesian neural networks
Title | ODE2VAE: Deep generative second order ODEs with Bayesian neural networks |
Authors | Cagatay Yildiz, Markus Heinonen, Harri Lahdesmaki |
Abstract | We present Ordinary Differential Equation Variational Auto-Encoder (ODE2VAE), a latent second order ODE model for high-dimensional sequential data. Leveraging the advances in deep generative models, ODE2VAE can simultaneously learn the embedding of high dimensional trajectories and infer arbitrarily complex continuous-time latent dynamics. Our model explicitly decomposes the latent space into momentum and position components and solves a second order ODE system, which is in contrast to recurrent neural network (RNN) based time series models and recently proposed black-box ODE techniques. In order to account for uncertainty, we propose probabilistic latent ODE dynamics parameterized by deep Bayesian neural networks. We demonstrate our approach on motion capture, image rotation, and bouncing balls datasets. We achieve state-of-the-art performance in long term motion prediction and imputation tasks. |
Tasks | Imputation, Motion Capture, motion prediction, Time Series |
Published | 2019-12-01 |
URL | http://papers.nips.cc/paper/9497-ode2vae-deep-generative-second-order-odes-with-bayesian-neural-networks |
http://papers.nips.cc/paper/9497-ode2vae-deep-generative-second-order-odes-with-bayesian-neural-networks.pdf | |
PWC | https://paperswithcode.com/paper/ode2vae-deep-generative-second-order-odes-1 |
Repo | https://github.com/cagatayyildiz/ODE2VAE |
Framework | tf |
As-projective-as-possible bias correction for illumination estimation algorithms
Title | As-projective-as-possible bias correction for illumination estimation algorithms |
Authors | Mahmoud Afifi, Abhijith Punnappurath, Graham Finlayson, Michael S. Brown |
Abstract | Illumination estimation is the key routine in a camera’s onboard auto-white-balance (AWB) function. Illumination estimation algorithms estimate the color of the scene’s illumination from an image in the form of an R,G,B vector in the sensor’s raw-RGB color space. While learning-based methods have demonstrated impressive performance for illumination estimation, cameras still rely on simple statistical-based algorithms that are less accurate but capable of executing quickly on the camera’s hardware. An effective strategy to improve the accuracy of these fast statistical-based algorithms is to apply a post-estimate bias correction function to transform the estimated R,G,B vector such that it lies closer to the correct solution. Recent work by Finlayson, Interface Focus, 2018 showed that a bias correction function can be formulated as a projective transform because the magnitude of the R,G,B illumination vector does not matter to the AWB procedure. This paper builds on this finding and shows that further improvements can be obtained by using an as-projective-as-possible (APAP) projective transform that locally adapts the projective transform to the input R,G,B vector. We demonstrate the effectiveness of the proposed APAP bias correction on several well-known statistical illumination estimation methods. We also describe a fast lookup method that allows the APAP transform to be performed with only a few lookup operations. |
Tasks | |
Published | 2019-01-01 |
URL | https://www.osapublishing.org/josaa/abstract.cfm?uri=josaa-36-1-71 |
http://www.cse.yorku.ca/~mbrown/pdf/josa_2019_afifi.pdf | |
PWC | https://paperswithcode.com/paper/as-projective-as-possible-bias-correction-for |
Repo | https://github.com/mahmoudnafifi/APAP-bias-correction-for-illumination-estimation-methods |
Framework | none |
GraphDTA: prediction of drug–target binding affinity using graph convolutional networks
Title | GraphDTA: prediction of drug–target binding affinity using graph convolutional networks |
Authors | Thin Nguyen, Hang Le, Svetha Venkatesh |
Abstract | While the development of new drugs is costly, time consuming, and often accompanied with safety issues, drug repurposing, where old drugs with established safety are used for medical conditions other than originally developed, is an attractive alternative. Then, how the old drugs work on new targets becomes a crucial part of drug repurposing and gains much of interest. Several statistical and machine learning models have been proposed to estimate drug–target binding affinity and deep learning approaches have been shown to be among state-of-the-art methods. However, drugs and targets in these models were commonly represented in 1D strings, regardless the fact that molecules are by nature formed by the chemical bonding of atoms. In this work, we propose GraphDTA to capture the structural information of drugs, possibly enhancing the predictive power of the affinity. In particular, unlike competing methods, drugs are represented as graphs and graph convolutional networks are used to learn drug–target binding affinity. We trial our method on two benchmark drug–target binding affinity datasets and compare the performance with state-of-the-art models in the field. The results show that our proposed method can not only predict the affinity better than non-deep learning models, but also outperform competing deep learning approaches. This demonstrates the practical advantages of graph-based representation for molecules in providing accurate prediction of drug–target binding affinity. The application may also include any recommendation systems where either or both of the user- and product-like sides can be represented in graphs. |
Tasks | Recommendation Systems |
Published | 2019-07-02 |
URL | https://www.biorxiv.org/content/10.1101/684662v3 |
https://www.biorxiv.org/content/biorxiv/early/2019/07/02/684662.full.pdf | |
PWC | https://paperswithcode.com/paper/graphdta-prediction-of-drugtarget-binding |
Repo | https://github.com/thinng/GraphDTA |
Framework | pytorch |
Event Detection with Trigger-Aware Lattice Neural Network
Title | Event Detection with Trigger-Aware Lattice Neural Network |
Authors | Ning Ding, Ziran Li, Zhiyuan Liu, Haitao Zheng, Zibo Lin |
Abstract | Event detection (ED) aims to locate trigger words in raw text and then classify them into correct event types. In this task, neural net- work based models became mainstream in re- cent years. However, two problems arise when it comes to languages without natural delim- iters, such as Chinese. First, word-based mod- els severely suffer from the problem of word- trigger mismatch, limiting the performance of the methods. In addition, even if trigger words could be accurately located, the ambi- guity of polysemy of triggers could still af- fect the trigger classification stage. To ad- dress the two issues simultaneously, we pro- pose the Trigger-aware Lattice Neural Net- work (TLNN). (1) The framework dynami- cally incorporates word and character informa- tion so that the trigger-word mismatch issue can be avoided. (2) Moreover, for polysemous characters and words, we model all senses of them with the help of an external linguistic knowledge base, so as to alleviate the prob- lem of ambiguous triggers. Experiments on two benchmark datasets show that our model could effectively tackle the two issues and outperforms previous state-of-the-art methods significantly, giving the best results. The source code of this paper can be obtained from https://github.com/thunlp/TLNN. |
Tasks | |
Published | 2019-11-01 |
URL | https://www.aclweb.org/anthology/D19-1033/ |
https://www.aclweb.org/anthology/D19-1033 | |
PWC | https://paperswithcode.com/paper/event-detection-with-trigger-aware-lattice |
Repo | https://github.com/thunlp/TLNN |
Framework | pytorch |
Actions Speak Louder than Goals:Valuing Player Actions in Soccer
Title | Actions Speak Louder than Goals:Valuing Player Actions in Soccer |
Authors | Tom Decroos, Lotte Bransen, Jan Van Haaren, Jesse Davis |
Abstract | Assessing the impact of the individual actions performed by soccerplayers during games is a crucial aspect of the player recruitmentprocess. Unfortunately, most traditional metrics fall short in ad-dressing this task as they either focus on rare actions like shotsand goals alone or fail to account for the context in which theactions occurred. This paper introduces (1) a new language for de-scribing individual player actions on the pitch and (2) a frameworkfor valuing any type of player action based on its impact on thegame outcome while accounting for the context in which the actionhappened. By aggregating soccer players’ action values, their totaloffensive and defensive contributions to their team can be quan-tified. We show how our approach considers relevant contextualinformation that traditional player evaluation metrics ignore andpresent a number of use cases related to scouting and playing stylecharacterization in the 2016/2017 and 2017/2018 seasons in Europe’stop competitions. |
Tasks | Football Action Valuation |
Published | 2019-07-10 |
URL | https://www.researchgate.net/publication/323302738_Actions_Speak_Louder_Than_Goals_Valuing_Player_Actions_in_Soccer |
https://arxiv.org/pdf/1802.07127.pdf | |
PWC | https://paperswithcode.com/paper/actions-speak-louder-than-goalsvaluing-player |
Repo | https://github.com/ML-KULeuven/socceraction |
Framework | none |
Unsupervised Discovery of Temporal Structure in Noisy Data with Dynamical Components Analysis
Title | Unsupervised Discovery of Temporal Structure in Noisy Data with Dynamical Components Analysis |
Authors | David Clark, Jesse Livezey, Kristofer Bouchard |
Abstract | Linear dimensionality reduction methods are commonly used to extract low-dimensional structure from high-dimensional data. However, popular methods disregard temporal structure, rendering them prone to extracting noise rather than meaningful dynamics when applied to time series data. At the same time, many successful unsupervised learning methods for temporal, sequential and spatial data extract features which are predictive of their surrounding context. Combining these approaches, we introduce Dynamical Components Analysis (DCA), a linear dimensionality reduction method which discovers a subspace of high-dimensional time series data with maximal predictive information, defined as the mutual information between the past and future. We test DCA on synthetic examples and demonstrate its superior ability to extract dynamical structure compared to commonly used linear methods. We also apply DCA to several real-world datasets, showing that the dimensions extracted by DCA are more useful than those extracted by other methods for predicting future states and decoding auxiliary variables. Overall, DCA robustly extracts dynamical structure in noisy, high-dimensional data while retaining the computational efficiency and geometric interpretability of linear dimensionality reduction methods. |
Tasks | Dimensionality Reduction, Time Series |
Published | 2019-12-01 |
URL | http://papers.nips.cc/paper/9574-unsupervised-discovery-of-temporal-structure-in-noisy-data-with-dynamical-components-analysis |
http://papers.nips.cc/paper/9574-unsupervised-discovery-of-temporal-structure-in-noisy-data-with-dynamical-components-analysis.pdf | |
PWC | https://paperswithcode.com/paper/unsupervised-discovery-of-temporal-structure-1 |
Repo | https://github.com/BouchardLab/DynamicalComponentsAnalysis |
Framework | none |
Joint Acne Image Grading and Counting via Label Distribution Learning
Title | Joint Acne Image Grading and Counting via Label Distribution Learning |
Authors | Xiaoping Wu, Ni Wen, Jie Liang, Yu-Kun Lai, Dongyu She, Ming-Ming Cheng, Jufeng Yang |
Abstract | Accurate grading of skin disease severity plays a crucial role in precise treatment for patients. Acne vulgaris, the most common skin disease in adolescence, can be graded by evidence-based lesion counting as well as experience-based global estimation in the medical field. However, due to the appearance similarity of acne with close severity, it is challenging to count and grade acne accurately. In this paper, we address the problem of acne image analysis via Label Distribution Learning (LDL) considering the ambiguous information among acne severity. Based on the professional grading criterion, we generate two acne label distributions considering the relationship between the similar number of lesions and severity of acne, respectively. We also propose a unified framework for joint acne image grading and counting, which is optimized by the multi-task learning loss. In addition, we further build the ACNE04 dataset with annotations of acne severity and lesion number of each image for evaluation. Experiments demonstrate that our proposed framework performs favorably against state-of-the-art methods. We make the code and dataset publicly available at https://github.com/xpwu95/ldl. |
Tasks | Multi-Task Learning |
Published | 2019-10-01 |
URL | http://openaccess.thecvf.com/content_ICCV_2019/html/Wu_Joint_Acne_Image_Grading_and_Counting_via_Label_Distribution_Learning_ICCV_2019_paper.html |
http://openaccess.thecvf.com/content_ICCV_2019/papers/Wu_Joint_Acne_Image_Grading_and_Counting_via_Label_Distribution_Learning_ICCV_2019_paper.pdf | |
PWC | https://paperswithcode.com/paper/joint-acne-image-grading-and-counting-via |
Repo | https://github.com/xpwu95/ldl |
Framework | pytorch |
CNN^{2}: Viewpoint Generalization via a Binocular Vision
Title | CNN^{2}: Viewpoint Generalization via a Binocular Vision |
Authors | Wei-Da Chen, Shan-Hung (Brandon) Wu |
Abstract | The Convolutional Neural Networks (CNNs) have laid the foundation for many techniques in various applications. Despite achieving remarkable performance in some tasks, the 3D viewpoint generalizability of CNNs is still far behind humans visual capabilities. Although recent efforts, such as the Capsule Networks, have been made to address this issue, these new models are either hard to train and/or incompatible with existing CNN-based techniques specialized for different applications. Observing that humans use binocular vision to understand the world, we study in this paper whether the 3D viewpoint generalizability of CNNs can be achieved via a binocular vision. We propose CNN^{2}, a CNN that takes two images as input, which resembles the process of an object being viewed from the left eye and the right eye. CNN^{2} uses novel augmentation, pooling, and convolutional layers to learn a sense of three-dimensionality in a recursive manner. Empirical evaluation shows that CNN^{2} has improved viewpoint generalizability compared to vanilla CNNs. Furthermore, CNN^{2} is easy to implement and train, and is compatible with existing CNN-based specialized techniques for different applications. |
Tasks | |
Published | 2019-12-01 |
URL | http://papers.nips.cc/paper/8473-cnn2-viewpoint-generalization-via-a-binocular-vision |
http://papers.nips.cc/paper/8473-cnn2-viewpoint-generalization-via-a-binocular-vision.pdf | |
PWC | https://paperswithcode.com/paper/cnn2-viewpoint-generalization-via-a-binocular |
Repo | https://github.com/wdchenxyz/CNN2 |
Framework | tf |
Using Natural Language Relations between Answer Choices for Machine Comprehension
Title | Using Natural Language Relations between Answer Choices for Machine Comprehension |
Authors | Rajkumar Pujari, Dan Goldwasser |
Abstract | While evaluating an answer choice for Reading Comprehension task, other answer choices available for the question and the answers of related questions about the same paragraph often provide valuable information. In this paper, we propose a method to leverage the natural language relations between the answer choices, such as entailment and contradiction, to improve the performance of machine comprehension. We use a stand-alone question answering (QA) system to perform QA task and a Natural Language Inference (NLI) system to identify the relations between the choice pairs. Then we perform inference using an Integer Linear Programming (ILP)-based relational framework to re-evaluate the decisions made by the standalone QA system in light of the relations identified by the NLI system. We also propose a multitask learning model that learns both the tasks jointly. |
Tasks | Natural Language Inference, Question Answering, Reading Comprehension |
Published | 2019-06-01 |
URL | https://www.aclweb.org/anthology/N19-1403/ |
https://www.aclweb.org/anthology/N19-1403 | |
PWC | https://paperswithcode.com/paper/using-natural-language-relations-between |
Repo | https://github.com/pujari-rajkumar/naacl2019 |
Framework | none |
Scalable Collapsed Inference for High-Dimensional Topic Models
Title | Scalable Collapsed Inference for High-Dimensional Topic Models |
Authors | Rashidul Islam, James Foulds |
Abstract | The bigger the corpus, the more topics it can potentially support. To truly make full use of massive text corpora, a topic model inference algorithm must therefore scale efficiently in 1) documents and 2) topics, while 3) achieving accurate inference. Previous methods have achieved two out of three of these criteria simultaneously, but never all three at once. In this paper, we develop an online inference algorithm for topic models which leverages stochasticity to scale well in the number of documents, sparsity to scale well in the number of topics, and which operates in the collapsed representation of the topic model for improved accuracy and run-time performance. We use a Monte Carlo inner loop in the online setting to approximate the collapsed variational Bayes updates in a sparse and efficient way, which we accomplish via the MetropolisHastings Walker method. We showcase our algorithm on LDA and the recently proposed mixed membership skip-gram topic model. Our method requires only amortized $O(k_{d})$ computation per word token instead of $O(K)$ operations, where the number of topics occurring for a particular document $k_{d}\ll$ the total number of topics in the corpus $K$, to converge to a high-quality solution. |
Tasks | Topic Models |
Published | 2019-06-01 |
URL | https://www.aclweb.org/anthology/N19-1291/ |
https://www.aclweb.org/anthology/N19-1291 | |
PWC | https://paperswithcode.com/paper/scalable-collapsed-inference-for-high |
Repo | https://github.com/dr97531/SparseSCVB0 |
Framework | none |
Multi-Criteria Dimensionality Reduction with Applications to Fairness
Title | Multi-Criteria Dimensionality Reduction with Applications to Fairness |
Authors | Uthaipon Tantipongpipat, Samira Samadi, Mohit Singh, Jamie H. Morgenstern, Santosh Vempala |
Abstract | Dimensionality reduction is a classical technique widely used for data analysis. One foundational instantiation is Principal Component Analysis (PCA), which minimizes the average reconstruction error. In this paper, we introduce the multi-criteria dimensionality reduction problem where we are given multiple objectives that need to be optimized simultaneously. As an application, our model captures several fairness criteria for dimensionality reduction such as the Fair-PCA problem introduced by Samadi et al. [NeurIPS18] and the Nash Social Welfare (NSW) problem. In the Fair-PCA problem, the input data is divided into k groups, and the goal is to find a single d-dimensional representation for all groups for which the maximum reconstruction error of any one group is minimized. In NSW the goal is to maximize the product of the individual variances of the groups achieved by the common low-dimensinal space. Our main result is an exact polynomial-time algorithm for the two-criteria dimensionality reduction problem when the two criteria are increasing concave functions. As an application of this result, we obtain a polynomial time algorithm for Fair-PCA for k=2 groups, resolving an open problem of Samadi et al.[NeurIPS18], and a polynomial time algorithm for NSW objective for k=2 groups. We also give approximation algorithms for k>2. Our technical contribution in the above results is to prove new low-rank properties of extreme point solutions to semi-definite programs. We conclude with the results of several experiments indicating improved performance and generalized application of our algorithm on real-world datasets. |
Tasks | Dimensionality Reduction |
Published | 2019-12-01 |
URL | http://papers.nips.cc/paper/9652-multi-criteria-dimensionality-reduction-with-applications-to-fairness |
http://papers.nips.cc/paper/9652-multi-criteria-dimensionality-reduction-with-applications-to-fairness.pdf | |
PWC | https://paperswithcode.com/paper/multi-criteria-dimensionality-reduction-with |
Repo | https://github.com/SDPforAll/multiCriteriaDimReduction |
Framework | none |