January 25, 2020

2051 words 10 mins read

Paper Group NANR 46

Paper Group NANR 46

Automatic Gender Identification and Reinflection in Arabic. Partners in Crime: Multi-view Sequential Inference for Movie Understanding. Dick-Preston and Morbo at SemEval-2019 Task 4: Transfer Learning for Hyperpartisan News Detection. Localized random projections challenge benchmarks for bio-plausible deep learning. DAGMapper: Learning to Map by Di …

Automatic Gender Identification and Reinflection in Arabic

Title Automatic Gender Identification and Reinflection in Arabic
Authors Nizar Habash, Houda Bouamor, Christine Chung
Abstract The impressive progress in many Natural Language Processing (NLP) applications has increased the awareness of some of the biases these NLP systems have with regards to gender identities. In this paper, we propose an approach to extend biased single-output gender-blind NLP systems with gender-specific alternative reinflections. We focus on Arabic, a gender-marking morphologically rich language, in the context of machine translation (MT) from English, and for first-person-singular constructions only. Our contributions are the development of a system-independent gender-awareness wrapper, and the building of a corpus for training and evaluating first-person-singular gender identification and reinflection in Arabic. Our results successfully demonstrate the viability of this approach with 8{%} relative increase in Bleu score for first-person-singular feminine, and 5.3{%} comparable increase for first-person-singular masculine on top of a state-of-the-art gender-blind MT system on a held-out test set.
Tasks Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3822/
PDF https://www.aclweb.org/anthology/W19-3822
PWC https://paperswithcode.com/paper/automatic-gender-identification-and
Repo
Framework

Partners in Crime: Multi-view Sequential Inference for Movie Understanding

Title Partners in Crime: Multi-view Sequential Inference for Movie Understanding
Authors Nikos Papasarantopoulos, Lea Frermann, Mirella Lapata, Shay B. Cohen
Abstract Multi-view learning algorithms are powerful representation learning tools, often exploited in the context of multimodal problems. However, for problems requiring inference at the token-level of a sequence (that is, a separate prediction must be made for every time step), it is often the case that single-view systems are used, or that more than one views are fused in a simple manner. We describe an incremental neural architecture paired with a novel training objective for incremental inference. The network operates on multi-view data. We demonstrate the effectiveness of our approach on the problem of predicting perpetrators in crime drama series, for which our model significantly outperforms previous work and strong baselines. Moreover, we introduce two tasks, crime case and speaker type tagging, that contribute to movie understanding and demonstrate the effectiveness of our model on them.
Tasks MULTI-VIEW LEARNING, Representation Learning
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1212/
PDF https://www.aclweb.org/anthology/D19-1212
PWC https://paperswithcode.com/paper/partners-in-crime-multi-view-sequential
Repo
Framework

Dick-Preston and Morbo at SemEval-2019 Task 4: Transfer Learning for Hyperpartisan News Detection

Title Dick-Preston and Morbo at SemEval-2019 Task 4: Transfer Learning for Hyperpartisan News Detection
Authors Tim Isbister, Fredrik Johansson
Abstract In a world of information operations, influence campaigns, and fake news, classification of news articles as following hyperpartisan argumentation or not is becoming increasingly important. We present a deep learning-based approach in which a pre-trained language model has been fine-tuned on domain-specific data and used for classification of news articles, as part of the SemEval-2019 task on hyperpartisan news detection. The suggested approach yields accuracy and F1-scores around 0.8 which places the best performing classifier among the top-5 systems in the competition.
Tasks Language Modelling, Transfer Learning
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2160/
PDF https://www.aclweb.org/anthology/S19-2160
PWC https://paperswithcode.com/paper/dick-preston-and-morbo-at-semeval-2019-task-4
Repo
Framework

Localized random projections challenge benchmarks for bio-plausible deep learning

Title Localized random projections challenge benchmarks for bio-plausible deep learning
Authors Bernd Illing, Wulfram Gerstner, Johanni Brea
Abstract Similar to models of brain-like computation, artificial deep neural networks rely on distributed coding, parallel processing and plastic synaptic weights. Training deep neural networks with the error-backpropagation algorithm, however, is considered bio-implausible. An appealing alternative to training deep neural networks is to use one or a few hidden layers with fixed random weights or trained with an unsupervised, local learning rule and train a single readout layer with a supervised, local learning rule. We find that a network of leaky-integrate-andfire neurons with fixed random, localized receptive fields in the hidden layer and spike timing dependent plasticity to train the readout layer achieves 98.1% test accuracy on MNIST, which is close to the optimal result achievable with error-backpropagation in non-convolutional networks of rate neurons with one hidden layer. To support the design choices of the spiking network, we systematically compare the classification performance of rate networks with a single hidden layer, where the weights of this layer are either random and fixed, trained with unsupervised Principal Component Analysis or Sparse Coding, or trained with the backpropagation algorithm. This comparison revealed, first, that unsupervised learning does not lead to better performance than fixed random projections for large hidden layers on digit classification (MNIST) and object recognition (CIFAR10); second, networks with random projections and localized receptive fields perform significantly better than networks with all-to-all connectivity and almost reach the performance of networks trained with the backpropagation algorithm. The performance of these simple random projection networks is comparable to most current models of bio-plausible deep learning and thus provides an interesting benchmark for future approaches.
Tasks Object Recognition
Published 2019-05-01
URL https://openreview.net/forum?id=SJeT_oRcY7
PDF https://openreview.net/pdf?id=SJeT_oRcY7
PWC https://paperswithcode.com/paper/localized-random-projections-challenge
Repo
Framework

DAGMapper: Learning to Map by Discovering Lane Topology

Title DAGMapper: Learning to Map by Discovering Lane Topology
Authors Namdar Homayounfar, Wei-Chiu Ma, Justin Liang, Xinyu Wu, Jack Fan, Raquel Urtasun
Abstract One of the fundamental challenges to scale self-driving is being able to create accurate high definition maps (HD maps) with low cost. Current attempts to automate this pro- cess typically focus on simple scenarios, estimate independent maps per frame or do not have the level of precision required by modern self driving vehicles. In contrast, in this paper we focus on drawing the lane boundaries of complex highways with many lanes that contain topology changes due to forks and merges. Towards this goal, we formulate the problem as inference in a directed acyclic graphical model (DAG), where the nodes of the graph encode geo- metric and topological properties of the local regions of the lane boundaries. Since we do not know a priori the topology of the lanes, we also infer the DAG topology (i.e., nodes and edges) for each region. We demonstrate the effectiveness of our approach on two major North American Highways in two different states and show high precision and recall as well as 89% correct topology.
Tasks
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Homayounfar_DAGMapper_Learning_to_Map_by_Discovering_Lane_Topology_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Homayounfar_DAGMapper_Learning_to_Map_by_Discovering_Lane_Topology_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/dagmapper-learning-to-map-by-discovering-lane
Repo
Framework

Studying Semantic Chain Shifts with Word2Vec: FOOD\textgreaterMEAT\textgreaterFLESH

Title Studying Semantic Chain Shifts with Word2Vec: FOOD\textgreaterMEAT\textgreaterFLESH
Authors Richard Zimmermann
Abstract Word2Vec models are used to study the semantic chain shift FOOD{\textgreater}MEAT{\textgreater}FLESH in the history of English, c. 1425-1925. The development stretches out over a long time, starting before 1500, and may possibly be continuing to this day. The semantic changes likely proceeded as a push chain.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4703/
PDF https://www.aclweb.org/anthology/W19-4703
PWC https://paperswithcode.com/paper/studying-semantic-chain-shifts-with-word2vec
Repo
Framework

Q-Theory Representations are Logically Equivalent to Autosegmental Representations

Title Q-Theory Representations are Logically Equivalent to Autosegmental Representations
Authors Nick Danis, Adam Jardine
Abstract
Tasks
Published 2019-01-01
URL https://www.aclweb.org/anthology/W19-0104/
PDF https://www.aclweb.org/anthology/W19-0104
PWC https://paperswithcode.com/paper/q-theory-representations-are-logically
Repo
Framework

Segmentation and UR Acquisition with UR Constraints

Title Segmentation and UR Acquisition with UR Constraints
Authors Max Nelson
Abstract
Tasks
Published 2019-01-01
URL https://www.aclweb.org/anthology/W19-0107/
PDF https://www.aclweb.org/anthology/W19-0107
PWC https://paperswithcode.com/paper/segmentation-and-ur-acquisition-with-ur
Repo
Framework

Number detectors spontaneously emerge in a deep neural network designed for visual object recognition

Title Number detectors spontaneously emerge in a deep neural network designed for visual object recognition
Authors Khaled Nasr, Pooja Viswanathan, Andreas Nieder
Abstract Humans and animals have a “number sense,” an innate capability to intuitively assess the number of visual items in a set, its numerosity. This capability implies that mechanisms to extract numerosity indwell the brain’s visual system, which is primarily concerned with visual object recognition. Here, we show that network units tuned to abstract numerosity, and therefore reminiscent of real number neurons, spontaneously emerge in a biologically inspired deep neural network that was merely trained on visual object recognition. These numerosity-tuned units underlay the network’s number discrimination performance that showed all the characteristics of human and animal number discriminations as predicted by the WeberFechner law. These findings explain the spontaneous emergence of the number sense based on mechanisms inherent to the visual system.
Tasks Object Recognition
Published 2019-05-08
URL https://advances.sciencemag.org/content/5/5/eaav7903.full
PDF https://advances.sciencemag.org/content/advances/5/5/eaav7903.full.pdf%20
PWC https://paperswithcode.com/paper/number-detectors-spontaneously-emerge-in-a
Repo
Framework

Proceedings of the Second Workshop on Computational Models of Reference, Anaphora and Coreference

Title Proceedings of the Second Workshop on Computational Models of Reference, Anaphora and Coreference
Authors
Abstract
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-2800/
PDF https://www.aclweb.org/anthology/W19-2800
PWC https://paperswithcode.com/paper/proceedings-of-the-second-workshop-on-15
Repo
Framework

Identifying Participation of Individual Verbs or VerbNet Classes in the Causative Alternation

Title Identifying Participation of Individual Verbs or VerbNet Classes in the Causative Alternation
Authors Esther Seyffarth
Abstract
Tasks Question Answering
Published 2019-01-01
URL https://www.aclweb.org/anthology/W19-0115/
PDF https://www.aclweb.org/anthology/W19-0115
PWC https://paperswithcode.com/paper/identifying-participation-of-individual-verbs
Repo
Framework

Jabberwocky Parsing: Dependency Parsing with Lexical Noise

Title Jabberwocky Parsing: Dependency Parsing with Lexical Noise
Authors Jungo Kasai, Robert Frank
Abstract
Tasks Dependency Parsing, Word Embeddings
Published 2019-01-01
URL https://www.aclweb.org/anthology/W19-0112/
PDF https://www.aclweb.org/anthology/W19-0112
PWC https://paperswithcode.com/paper/jabberwocky-parsing-dependency-parsing-with
Repo
Framework

Versatile Multiple Choice Learning and Its Application to Vision Computing

Title Versatile Multiple Choice Learning and Its Application to Vision Computing
Authors Kai Tian, Yi Xu, Shuigeng Zhou, Jihong Guan
Abstract Most existing ensemble methods aim to train the underlying embedded models independently and simply aggregate their final outputs via averaging or weighted voting. As many prediction tasks contain uncertainty, most of these ensemble methods just reduce variance of the predictions without considering the collaborations among the ensembles. Different from these ensemble methods, multiple choice learning (MCL) methods exploit the cooperation among all the embedded models to generate multiple diverse hypotheses. In this paper, a new MCL method, called vMCL (the abbreviation of versatile Multiple Choice Learning), is developed to extend the application scenarios of MCL methods by ensembling deep neural networks. Our vMCL method keeps the advantage of existing MCL methods while overcoming their major drawback, thus achieves better performance. The novelty of our vMCL lies in three aspects: (1) a choice network is designed to learn the confidence level of each specialist which can provide the best prediction base on multiple hypotheses; (2) a hinge loss is introduced to alleviate the overconfidence issue in MCL settings; (3) Easy to be implemented and can be trained in an end-to-end manner, which is a very attractive feature for many real-world applications. Experiments on image classification and image segmentation task show that vMCL outperforms the existing state-of-the-art MCL methods.
Tasks Image Classification, Semantic Segmentation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Tian_Versatile_Multiple_Choice_Learning_and_Its_Application_to_Vision_Computing_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Tian_Versatile_Multiple_Choice_Learning_and_Its_Application_to_Vision_Computing_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/versatile-multiple-choice-learning-and-its
Repo
Framework

Oracle-Efficient Algorithms for Online Linear Optimization with Bandit Feedback

Title Oracle-Efficient Algorithms for Online Linear Optimization with Bandit Feedback
Authors Shinji Ito, Daisuke Hatano, Hanna Sumita, Kei Takemura, Takuro Fukunaga, Naonori Kakimura, Ken-Ichi Kawarabayashi
Abstract We propose computationally efficient algorithms for \textit{online linear optimization with bandit feedback}, in which a player chooses an \textit{action vector} from a given (possibly infinite) set $\mathcal{A} \subseteq \mathbb{R}^d$, and then suffers a loss that can be expressed as a linear function in action vectors. Although existing algorithms achieve an optimal regret bound of $\tilde{O}(\sqrt{T})$ for $T$ rounds (ignoring factors of $\mathrm{poly} (d, \log T)$), computationally efficient ways of implementing them have not yet been specified, in particular when $\mathcal{A}$ is not bounded by a polynomial size in $d$. A standard way to pursue computational efficiency is to assume that we have an efficient algorithm referred to as \textit{oracle} that solves (offline) linear optimization problems over $\mathcal{A}$. Under this assumption, the computational efficiency of a bandit algorithm can then be measured in terms of \textit{oracle complexity}, i.e., the number of oracle calls. Our contribution is to propose algorithms that offer optimal regret bounds of $\tilde{O}(\sqrt{T})$ as well as low oracle complexity for both \textit{non-stochastic settings} and \textit{stochastic settings}. Our algorithm for non-stochastic settings has an oracle complexity of $\tilde{O}( T )$ and is the first algorithm that achieves both a regret bound of $\tilde{O}( \sqrt{T} )$ and an oracle complexity of $\tilde{O} ( \mathrm{poly} ( T ) )$, given only linear optimization oracles. Our algorithm for stochastic settings calls the oracle only $O( \mathrm{poly} (d, \log T))$ times, which is smaller than the current best oracle complexity of $O( T )$ if $T$ is sufficiently large.
Tasks
Published 2019-12-01
URL http://papers.nips.cc/paper/9244-oracle-efficient-algorithms-for-online-linear-optimization-with-bandit-feedback
PDF http://papers.nips.cc/paper/9244-oracle-efficient-algorithms-for-online-linear-optimization-with-bandit-feedback.pdf
PWC https://paperswithcode.com/paper/oracle-efficient-algorithms-for-online-linear
Repo
Framework

Experiments on Non-native Speech Assessment and its Consistency

Title Experiments on Non-native Speech Assessment and its Consistency
Authors Ziwei Zhou, Sowmya Vajjala, Seyed Vahid Mirnezami
Abstract
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-6309/
PDF https://www.aclweb.org/anthology/W19-6309
PWC https://paperswithcode.com/paper/experiments-on-non-native-speech-assessment
Repo
Framework
comments powered by Disqus