January 24, 2020

2023 words 10 mins read

Paper Group NANR 135

Paper Group NANR 135

The AFRL WMT19 Systems: Old Favorites and New Tricks. Proceedings of the Human-Informed Translation and Interpreting Technology Workshop (HiT-IT 2019). Proceedings of the First Workshop on NLP for Conversational AI. SIMILE: Introducing Sequential Information towards More Effective Imitation Learning. Seeing more than whitespace — Tokenisation and …

The AFRL WMT19 Systems: Old Favorites and New Tricks

Title The AFRL WMT19 Systems: Old Favorites and New Tricks
Authors Jeremy Gwinnup, Grant Erdmann, Tim Anderson
Abstract This paper describes the Air Force Research Laboratory (AFRL) machine translation systems and the improvements that were developed during the WMT19 evaluation campaign. This year, we refine our approach to training popular neural machine translation toolkits, experiment with a new domain adaptation technique and again measure improvements in performance on the Russian{–}English language pair.
Tasks Domain Adaptation, Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5318/
PDF https://www.aclweb.org/anthology/W19-5318
PWC https://paperswithcode.com/paper/the-afrl-wmt19-systems-old-favorites-and-new
Repo
Framework

Proceedings of the Human-Informed Translation and Interpreting Technology Workshop (HiT-IT 2019)

Title Proceedings of the Human-Informed Translation and Interpreting Technology Workshop (HiT-IT 2019)
Authors
Abstract
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-8700/
PDF https://www.aclweb.org/anthology/W19-8700
PWC https://paperswithcode.com/paper/proceedings-of-the-human-informed-translation
Repo
Framework

Proceedings of the First Workshop on NLP for Conversational AI

Title Proceedings of the First Workshop on NLP for Conversational AI
Authors
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4100/
PDF https://www.aclweb.org/anthology/W19-4100
PWC https://paperswithcode.com/paper/proceedings-of-the-first-workshop-on-nlp-for
Repo
Framework

SIMILE: Introducing Sequential Information towards More Effective Imitation Learning

Title SIMILE: Introducing Sequential Information towards More Effective Imitation Learning
Authors Yutong Bai, Lingxi Xie
Abstract Reinforcement learning (RL) is a metaheuristic aiming at teaching an agent to interact with an environment and maximizing the reward in a complex task. RL algorithms often encounter the difficulty in defining a reward function in a sparse solution space. Imitation learning (IL) deals with this issue by providing a few expert demonstrations, and then either mimicking the expert’s behavior (behavioral cloning, BC) or recovering the reward function by assuming the optimality of the expert (inverse reinforcement learning, IRL). Conventional IL approaches formulate the agent policy by mapping one single state to a distribution over actions, which did not consider sequential information. This strategy can be less accurate especially in IL, a weakly supervised learning environment, especially when the number of expert demonstrations is limited. This paper presents an effective approach named Sequential IMItation LEarning (SIMILE). The core idea is to introduce sequential information, so that an agent can refer to both the current state and past state-action pairs to make a decision. We formulate our approach into a recurrent model, and instantiate it using LSTM so as to fuse both long-term and short-term information. SIMILE is a generalized IL framework which is easily applied to BL and IRL, two major types of IL algorithms. Experiments are performed on several robot controlling tasks in OpenAI Gym. SIMILE not only achieves performance gain over the baseline approaches, but also enjoys the benefit of faster convergence and better stability of testing performance. These advantages verify a higher learning efficiency of SIMILE, and implies its potential applications in real-world scenarios, i.e., when the agent-environment interaction is more difficult and/or expensive.
Tasks Imitation Learning
Published 2019-05-01
URL https://openreview.net/forum?id=Hyghb2Rct7
PDF https://openreview.net/pdf?id=Hyghb2Rct7
PWC https://paperswithcode.com/paper/simile-introducing-sequential-information
Repo
Framework

Seeing more than whitespace — Tokenisation and disambiguation in a North S'ami grammar checker

Title Seeing more than whitespace — Tokenisation and disambiguation in a North S'ami grammar checker
Authors Linda Wiechetek, Sjur N{\o}rsteb{\o} Moshagen, Kevin Brubeck Unhammer Tokenization, {disambiguation of potential compounds in North S{'a}mi grammar checking}
Abstract
Tasks
Published 2019-02-01
URL https://www.aclweb.org/anthology/W19-6007/
PDF https://www.aclweb.org/anthology/W19-6007
PWC https://paperswithcode.com/paper/seeing-more-than-whitespace-tokenisation-and
Repo
Framework

Pix2Scene: Learning Implicit 3D Representations from Images

Title Pix2Scene: Learning Implicit 3D Representations from Images
Authors Sai Rajeswar, Fahim Mannan, Florian Golemo, David Vazquez, Derek Nowrouzezahrai, Aaron Courville
Abstract Modelling 3D scenes from 2D images is a long-standing problem in computer vision with implications in, e.g., simulation and robotics. We propose pix2scene, a deep generative-based approach that implicitly models the geometric properties of a scene from images. Our method learns the depth and orientation of scene points visible in images. Our model can then predict the structure of a scene from various, previously unseen view points. It relies on a bi-directional adversarial learning mechanism to generate scene representations from a latent code, inferring the 3D representation of the underlying scene geometry. We showcase a novel differentiable renderer to train the 3D model in an end-to-end fashion, using only images. We demonstrate the generative ability of our model qualitatively on both a custom dataset and on ShapeNet. Finally, we evaluate the effectiveness of the learned 3D scene representation in supporting a 3D spatial reasoning.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=BJeem3C9F7
PDF https://openreview.net/pdf?id=BJeem3C9F7
PWC https://paperswithcode.com/paper/pix2scene-learning-implicit-3d
Repo
Framework

L3-Net: Towards Learning Based LiDAR Localization for Autonomous Driving

Title L3-Net: Towards Learning Based LiDAR Localization for Autonomous Driving
Authors Weixin Lu, Yao Zhou, Guowei Wan, Shenhua Hou, Shiyu Song
Abstract We present L3-Net - a novel learning-based LiDAR localization system that achieves centimeter-level localization accuracy, comparable to prior state-of-the-art systems with hand-crafted pipelines. Rather than relying on these hand-crafted modules, we innovatively implement the use of various deep neural network structures to establish a learning-based approach. L3-Net learns local descriptors specifically optimized for matching in different real-world driving scenarios. 3D convolutions over a cost volume built in the solution space significantly boosts the localization accuracy. RNNs are demonstrated to be effective in modeling the vehicle’s dynamics, yielding better temporal smoothness and accuracy. We comprehensively validate the effectiveness of our approach using freshly collected datasets. Multiple trials of repetitive data collection over the same road and areas make our dataset ideal for testing localization systems. The SunnyvaleBigLoop sequences, with a year’s time interval between the collected mapping and testing data, made it quite challenging, but the low localization error of our method in these datasets demonstrates its maturity for real industrial implementation.
Tasks Autonomous Driving
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Lu_L3-Net_Towards_Learning_Based_LiDAR_Localization_for_Autonomous_Driving_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Lu_L3-Net_Towards_Learning_Based_LiDAR_Localization_for_Autonomous_Driving_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/l3-net-towards-learning-based-lidar
Repo
Framework

Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1)

Title Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1)
Authors
Abstract
Tasks Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5300/
PDF https://www.aclweb.org/anthology/W19-5300
PWC https://paperswithcode.com/paper/proceedings-of-the-fourth-conference-on-2
Repo
Framework

Grammar Error Correction in Morphologically Rich Languages: The Case of Russian

Title Grammar Error Correction in Morphologically Rich Languages: The Case of Russian
Authors Alla Rozovskaya, Dan Roth
Abstract Until now, most of the research in grammar error correction focused on English, and the problem has hardly been explored for other languages. We address the task of correcting writing mistakes in morphologically rich languages, with a focus on Russian. We present a corrected and error-tagged corpus of Russian learner writing and develop models that make use of existing state-of-the-art methods that have been well studied for English. Although impressive results have recently been achieved for grammar error correction of non-native English writing, these results are limited to domains where plentiful training data are available. Because annotation is extremely costly, these approaches are not suitable for the majority of domains and languages. We thus focus on methods that use {``}minimal supervision{''}; that is, those that do not rely on large amounts of annotated training data, and show how existing minimal-supervision approaches extend to a highly inflectional language such as Russian. The results demonstrate that these methods are particularly useful for correcting mistakes in grammatical phenomena that involve rich morphology. |
Tasks
Published 2019-03-01
URL https://www.aclweb.org/anthology/Q19-1001/
PDF https://www.aclweb.org/anthology/Q19-1001
PWC https://paperswithcode.com/paper/grammar-error-correction-in-morphologically
Repo
Framework

From Shakespeare to Li-Bai: Adapting a Sonnet Model to Chinese Poetry

Title From Shakespeare to Li-Bai: Adapting a Sonnet Model to Chinese Poetry
Authors Zhuohan Xie, Jey Han Lau, Trevor Cohn
Abstract In this paper, we adapt Deep-speare, a joint neural network model for English sonnets, to Chinese poetry. We illustrate characteristics of Chinese quatrain and explain our architecture as well as training and generation procedure, which differs from Shakespeare sonnets in several aspects. We analyse the generated poetry and find that model works well for Chinese poetry, as it can: (1) generate coherent 4-line quatrains of different topics; and (2) capture rhyme automatically (to a certain extent).
Tasks
Published 2019-04-01
URL https://www.aclweb.org/anthology/U19-1002/
PDF https://www.aclweb.org/anthology/U19-1002
PWC https://paperswithcode.com/paper/from-shakespeare-to-li-bai-adapting-a-sonnet
Repo
Framework

On the Compositionality Prediction of Noun Phrases using Poincar'e Embeddings

Title On the Compositionality Prediction of Noun Phrases using Poincar'e Embeddings
Authors Abhik Jana, Dima Puzyrev, Alex Panchenko, er, Pawan Goyal, Chris Biemann, Animesh Mukherjee
Abstract The compositionality degree of multiword expressions indicates to what extent the meaning of a phrase can be derived from the meaning of its constituents and their grammatical relations. Prediction of (non)-compositionality is a task that has been frequently addressed with distributional semantic models. We introduce a novel technique to blend hierarchical information with distributional information for predicting compositionality. In particular, we use hypernymy information of the multiword and its constituents encoded in the form of the recently introduced Poincar{'e} embeddings in addition to the distributional information to detect compositionality for noun phrases. Using a weighted average of the distributional similarity and a Poincar{'e} similarity function, we obtain consistent and substantial, statistically significant improvement across three gold standard datasets over state-of-the-art models based on distributional information only. Unlike traditional approaches that solely use an unsupervised setting, we have also framed the problem as a supervised task, obtaining comparable improvements. Further, we publicly release our Poincar{'e} embeddings, which are trained on the output of handcrafted lexical-syntactic patterns on a large corpus.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1316/
PDF https://www.aclweb.org/anthology/P19-1316
PWC https://paperswithcode.com/paper/on-the-compositionality-prediction-of-noun-1
Repo
Framework

Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Student Research Workshop

Title Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Student Research Workshop
Authors
Abstract
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-3000/
PDF https://www.aclweb.org/anthology/N19-3000
PWC https://paperswithcode.com/paper/proceedings-of-the-2019-conference-of-the-2
Repo
Framework

Mixed Multi-Head Self-Attention for Neural Machine Translation

Title Mixed Multi-Head Self-Attention for Neural Machine Translation
Authors Hongyi Cui, Shohei Iida, Po-Hsuan Hung, Takehito Utsuro, Masaaki Nagata
Abstract Recently, the Transformer becomes a state-of-the-art architecture in the filed of neural machine translation (NMT). A key point of its high-performance is the multi-head self-attention which is supposed to allow the model to independently attend to information from different representation subspaces. However, there is no explicit mechanism to ensure that different attention heads indeed capture different features, and in practice, redundancy has occurred in multiple heads. In this paper, we argue that using the same global attention in multiple heads limits multi-head self-attention{'}s capacity for learning distinct features. In order to improve the expressiveness of multi-head self-attention, we propose a novel Mixed Multi-Head Self-Attention (MMA) which models not only global and local attention but also forward and backward attention in different attention heads. This enables the model to learn distinct representations explicitly among multiple heads. In our experiments on both WAT17 English-Japanese as well as IWSLT14 German-English translation task, we show that, without increasing the number of parameters, our models yield consistent and significant improvements (0.9 BLEU scores on average) over the strong Transformer baseline.
Tasks Machine Translation
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5622/
PDF https://www.aclweb.org/anthology/D19-5622
PWC https://paperswithcode.com/paper/mixed-multi-head-self-attention-for-neural
Repo
Framework

A Comparison on Fine-grained Pre-trained Embeddings for the WMT19Chinese-English News Translation Task

Title A Comparison on Fine-grained Pre-trained Embeddings for the WMT19Chinese-English News Translation Task
Authors Zhenhao Li, Lucia Specia
Abstract This paper describes our submission to the WMT 2019 Chinese-English (zh-en) news translation shared task. Our systems are based on RNN architectures with pre-trained embeddings which utilize character and sub-character information. We compare models with these different granularity levels using different evaluating metics. We find that a finer granularity embeddings can help the model according to character level evaluation and that the pre-trained embeddings can also be beneficial for model performance marginally when the training data is limited.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5324/
PDF https://www.aclweb.org/anthology/W19-5324
PWC https://paperswithcode.com/paper/a-comparison-on-fine-grained-pre-trained
Repo
Framework

Cognate Projection for Low-Resource Inflection Generation

Title Cognate Projection for Low-Resource Inflection Generation
Authors Bradley Hauer, Amir Ahmad Habibi, Yixing Luan, Rashed Rubby Riyadh, Grzegorz Kondrak
Abstract We propose cognate projection as a method of crosslingual transfer for inflection generation in the context of the SIGMORPHON 2019 Shared Task. The results on four language pairs show the method is effective when no low-resource training data is available.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4202/
PDF https://www.aclweb.org/anthology/W19-4202
PWC https://paperswithcode.com/paper/cognate-projection-for-low-resource
Repo
Framework
comments powered by Disqus