January 24, 2020

2231 words 11 mins read

Paper Group NANR 210

Paper Group NANR 210

Relevant and Informative Response Generation using Pointwise Mutual Information. Learning Relational Representations by Analogy using Hierarchical Siamese Networks. Abstract Graphs and Abstract Paths for Knowledge Graph Completion. Efficiency Metrics for Data-Driven Models: A Text Summarization Case Study. Synchronously Generating Two Languages wit …

Relevant and Informative Response Generation using Pointwise Mutual Information

Title Relevant and Informative Response Generation using Pointwise Mutual Information
Authors Junya Takayama, Yuki Arase
Abstract A sequence-to-sequence model tends to generate generic responses with little information for input utterances. To solve this problem, we propose a neural model that generates relevant and informative responses. Our model has simple architecture to enable easy application to existing neural dialogue models. Specifically, using positive pointwise mutual information, it first identifies keywords that frequently co-occur in responses given an utterance. Then, the model encourages the decoder to use the keywords for response generation. Experiment results demonstrate that our model successfully diversifies responses relative to previous models.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4115/
PDF https://www.aclweb.org/anthology/W19-4115
PWC https://paperswithcode.com/paper/relevant-and-informative-response-generation
Repo
Framework

Learning Relational Representations by Analogy using Hierarchical Siamese Networks

Title Learning Relational Representations by Analogy using Hierarchical Siamese Networks
Authors Gaetano Rossiello, Alfio Gliozzo, Robert Farrell, Nicolas Fauceglia, Michael Glass
Abstract We address relation extraction as an analogy problem by proposing a novel approach to learn representations of relations expressed by their textual mentions. In our assumption, if two pairs of entities belong to the same relation, then those two pairs are analogous. Following this idea, we collect a large set of analogous pairs by matching triples in knowledge bases with web-scale corpora through distant supervision. We leverage this dataset to train a hierarchical siamese network in order to learn entity-entity embeddings which encode relational information through the different linguistic paraphrasing expressing the same relation. We evaluate our model in a one-shot learning task by showing a promising generalization capability in order to classify unseen relation types, which makes this approach suitable to perform automatic knowledge base population with minimal supervision. Moreover, the model can be used to generate pre-trained embeddings which provide a valuable signal when integrated into an existing neural-based model by outperforming the state-of-the-art methods on a downstream relation extraction task.
Tasks Entity Embeddings, Knowledge Base Population, One-Shot Learning, Relation Extraction
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-1327/
PDF https://www.aclweb.org/anthology/N19-1327
PWC https://paperswithcode.com/paper/learning-relational-representations-by
Repo
Framework

Abstract Graphs and Abstract Paths for Knowledge Graph Completion

Title Abstract Graphs and Abstract Paths for Knowledge Graph Completion
Authors Vivi Nastase, Bhushan Kotnis
Abstract Knowledge graphs, which provide numerous facts in a machine-friendly format, are incomplete. Information that we induce from such graphs {–} e.g. entity embeddings, relation representations or patterns {–} will be affected by the imbalance in the information captured in the graph {–} by biasing representations, or causing us to miss potential patterns. To partially compensate for this situation we describe a method for representing knowledge graphs that capture an intensional representation of the original extensional information. This representation is very compact, and it abstracts away from individual links, allowing us to find better path candidates, as shown by the results of link prediction using this information.
Tasks Entity Embeddings, Knowledge Graph Completion, Knowledge Graphs, Link Prediction
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-1016/
PDF https://www.aclweb.org/anthology/S19-1016
PWC https://paperswithcode.com/paper/abstract-graphs-and-abstract-paths-for
Repo
Framework

Efficiency Metrics for Data-Driven Models: A Text Summarization Case Study

Title Efficiency Metrics for Data-Driven Models: A Text Summarization Case Study
Authors Erion {\c{C}}ano, Ond{\v{r}}ej Bojar
Abstract Using data-driven models for solving text summarization or similar tasks has become very common in the last years. Yet most of the studies report basic accuracy scores only, and nothing is known about the ability of the proposed models to improve when trained on more data. In this paper, we define and propose three data efficiency metrics: data score efficiency, data time deficiency and overall data efficiency. We also propose a simple scheme that uses those metrics and apply it for a more comprehensive evaluation of popular methods on text summarization and title generation tasks. For the latter task, we process and release a huge collection of 35 million abstract-title pairs from scientific articles. Our results reveal that among the tested models, the Transformer is the most efficient on both tasks.
Tasks Text Summarization
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-8630/
PDF https://www.aclweb.org/anthology/W19-8630
PWC https://paperswithcode.com/paper/efficiency-metrics-for-data-driven-models-a-1
Repo
Framework

Synchronously Generating Two Languages with Interactive Decoding

Title Synchronously Generating Two Languages with Interactive Decoding
Authors Yining Wang, Jiajun Zhang, Long Zhou, Yuchen Liu, Chengqing Zong
Abstract In this paper, we introduce a novel interactive approach to translate a source language into two different languages simultaneously and interactively. Specifically, the generation of one language relies on not only previously generated outputs by itself, but also the outputs predicted in the other language. Experimental results on IWSLT and WMT datasets demonstrate that our method can obtain significant improvements over both conventional Neural Machine Translation (NMT) model and multilingual NMT model.
Tasks Machine Translation
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1330/
PDF https://www.aclweb.org/anthology/D19-1330
PWC https://paperswithcode.com/paper/synchronously-generating-two-languages-with
Repo
Framework

NRC Parallel Corpus Filtering System for WMT 2019

Title NRC Parallel Corpus Filtering System for WMT 2019
Authors Gabriel Bernier-Colborne, Chi-kiu Lo
Abstract We describe the National Research Council Canada team{'}s submissions to the parallel corpus filtering task at the Fourth Conference on Machine Translation.
Tasks Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5434/
PDF https://www.aclweb.org/anthology/W19-5434
PWC https://paperswithcode.com/paper/nrc-parallel-corpus-filtering-system-for-wmt
Repo
Framework

Visceral Machines: Reinforcement Learning with Intrinsic Physiological Rewards

Title Visceral Machines: Reinforcement Learning with Intrinsic Physiological Rewards
Authors Daniel McDuff, Ashish Kapoor
Abstract The human autonomic nervous system has evolved over millions of years and is essential for survival and responding to threats. As people learn to navigate the world, ``fight or flight’’ responses provide intrinsic feedback about the potential consequence of action choices (e.g., becoming nervous when close to a cliff edge or driving fast around a bend.) Physiological changes are correlated with these biological preparations to protect one-self from danger. We present a novel approach to reinforcement learning that leverages a task-independent intrinsic reward function trained on peripheral pulse measurements that are correlated with human autonomic nervous system responses. Our hypothesis is that such reward functions can circumvent the challenges associated with sparse and skewed rewards in reinforcement learning settings and can help improve sample efficiency. We test this in a simulated driving environment and show that it can increase the speed of learning and reduce the number of collisions during the learning stage. |
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=SyNvti09KQ
PDF https://openreview.net/pdf?id=SyNvti09KQ
PWC https://paperswithcode.com/paper/visceral-machines-reinforcement-learning-with-1
Repo
Framework

COCO-GAN: Conditional Coordinate Generative Adversarial Network

Title COCO-GAN: Conditional Coordinate Generative Adversarial Network
Authors Chieh Hubert Lin, Chia-Che Chang, Yu-Sheng Chen, Da-Cheng Juan, Wei Wei, Hwann-Tzong Chen
Abstract Recent advancements on Generative Adversarial Network (GAN) have inspired a wide range of works that generate synthetic images. However, the current processes have to generate an entire image at once, and therefore resolutions are limited by memory or computational constraints. In this work, we propose COnditional COordinate GAN (COCO-GAN), which generates a specific patch of an image conditioned on a spatial position rather than the entire image at a time. The generated patches are later combined together to form a globally coherent full-image. With this process, we show that the generated image can achieve competitive quality to state-of-the-arts and the generated patches are locally smooth between consecutive neighbors. One direct implication of the COCO-GAN is that it can be applied onto any coordinate systems including the cylindrical systems which makes it feasible for generating panorama images. The fact that the patch generation process is independent to each other inspires a wide range of new applications: firstly, “Patch-Inspired Image Generation” enables us to generate the entire image based on a single patch. Secondly, “Partial-Scene Generation” allows us to generate images within a customized target region. Finally, thanks to COCO-GAN’s patch generation and massive parallelism, which enables combining patches for generating a full-image with higher resolution than state-of-the-arts.
Tasks Image Generation, Scene Generation
Published 2019-05-01
URL https://openreview.net/forum?id=r14Aas09Y7
PDF https://openreview.net/pdf?id=r14Aas09Y7
PWC https://paperswithcode.com/paper/coco-gan-conditional-coordinate-generative
Repo
Framework

Proceedings of the Workshop on Extracting Structured Knowledge from Scientific Publications

Title Proceedings of the Workshop on Extracting Structured Knowledge from Scientific Publications
Authors
Abstract
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-2600/
PDF https://www.aclweb.org/anthology/W19-2600
PWC https://paperswithcode.com/paper/proceedings-of-the-workshop-on-extracting
Repo
Framework

Towards Annotating and Creating Summary Highlights at Sub-sentence Level

Title Towards Annotating and Creating Summary Highlights at Sub-sentence Level
Authors Kristjan Arumae, Parminder Bhatia, Fei Liu
Abstract Highlighting is a powerful tool to pick out important content and emphasize. Creating summary highlights at the sub-sentence level is particularly desirable, because sub-sentences are more concise than whole sentences. They are also better suited than individual words and phrases that can potentially lead to disfluent, fragmented summaries. In this paper we seek to generate summary highlights by annotating summary-worthy sub-sentences and teaching classifiers to do the same. We frame the task as jointly selecting important sentences and identifying a single most informative textual unit from each sentence. This formulation dramatically reduces the task complexity involved in sentence compression. Our study provides new benchmarks and baselines for generating highlights at the sub-sentence level.
Tasks Sentence Compression
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5408/
PDF https://www.aclweb.org/anthology/D19-5408
PWC https://paperswithcode.com/paper/towards-annotating-and-creating-summary
Repo
Framework

“Double-DIP”: Unsupervised Image Decomposition via Coupled Deep-Image-Priors

Title “Double-DIP”: Unsupervised Image Decomposition via Coupled Deep-Image-Priors
Authors Yosef Gandelsman, Assaf Shocher, Michal Irani
Abstract Many seemingly unrelated computer vision tasks can be viewed as a special case of image decomposition into separate layers. For example, image segmentation (separation into foreground and background layers); transparent layer separation (into reflection and transmission layers); Image dehazing (separation into a clear image and a haze map), and more. In this paper we propose a unified framework for unsupervised layer decomposition of a single image, based on coupled “Deep-image-Prior” (DIP) networks. It was shown [Ulyanov et al] that the structure of a single DIP generator network is sufficient to capture the low-level statistics of a single image. We show that coupling multiple such DIPs provides a powerful tool for decomposing images into their basic components, for a wide variety of applications. This capability stems from the fact that the internal statistics of a mixture of layers is more complex than the statistics of each of its individual components. We show the power of this approach for Image-Dehazing, Fg/Bg Segmentation, Watermark-Removal, Transparency Separation in images and video, and more. These capabilities are achieved in a totally unsupervised way, with no training examples other than the input image/video itself.
Tasks Image Dehazing, Semantic Segmentation, Transparency Separation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Gandelsman_Double-DIP_Unsupervised_Image_Decomposition_via_Coupled_Deep-Image-Priors_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Gandelsman_Double-DIP_Unsupervised_Image_Decomposition_via_Coupled_Deep-Image-Priors_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/double-dip-unsupervised-image-decomposition-1
Repo
Framework

Autoencoder-based Music Translation

Title Autoencoder-based Music Translation
Authors Noam Mor, Lior Wolf, Adam Polyak, Yaniv Taigman
Abstract We present a method for translating music across musical instruments and styles. This method is based on unsupervised training of a multi-domain wavenet autoencoder, with a shared encoder and a domain-independent latent space that is trained end-to-end on waveforms. Employing a diverse training dataset and large net capacity, the single encoder allows us to translate also from musical domains that were not seen during training. We evaluate our method on a dataset collected from professional musicians, and achieve convincing translations. We also study the properties of the obtained translation and demonstrate translating even from a whistle, potentially enabling the creation of instrumental music by untrained humans.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=HJGkisCcKm
PDF https://openreview.net/pdf?id=HJGkisCcKm
PWC https://paperswithcode.com/paper/autoencoder-based-music-translation
Repo
Framework

Team Ned Leeds at SemEval-2019 Task 4: Exploring Language Indicators of Hyperpartisan Reporting

Title Team Ned Leeds at SemEval-2019 Task 4: Exploring Language Indicators of Hyperpartisan Reporting
Authors Bozhidar Stevanoski, Sonja Gievska
Abstract This paper reports an experiment carried out to investigate the relevance of several syntactic, stylistic and pragmatic features on the task of distinguishing between mainstream and partisan news articles. The results of the evaluation of different feature sets and the extent to which various feature categories could affect the performance metrics are discussed and compared. Among different combinations of features and classifiers, Random Forest classifier using vector representations of the headline and the text of the report, with the inclusion of 8 readability scores and few stylistic features yielded best result, ranking our team at the 9th place at the SemEval 2019 Hyperpartisan News Detection challenge.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2179/
PDF https://www.aclweb.org/anthology/S19-2179
PWC https://paperswithcode.com/paper/team-ned-leeds-at-semeval-2019-task-4
Repo
Framework

Modeling Quantification and Scope in Abstract Meaning Representations

Title Modeling Quantification and Scope in Abstract Meaning Representations
Authors James Pustejovsky, Ken Lai, Nianwen Xue
Abstract In this paper, we propose an extension to Abstract Meaning Representations (AMRs) to encode scope information of quantifiers and negation, in a way that overcomes the semantic gaps of the schema while maintaining its cognitive simplicity. Specifically, we address three phenomena not previously part of the AMR specification: quantification, negation (generally), and modality. The resulting representation, which we call {}Uniform Meaning Representation{''} (UMR), adopts the predicative core of AMR and embeds it under a {}scope{''} graph when appropriate. UMR representations differ from other treatments of quantification and modal scope phenomena in two ways: (a) they are more transparent; and (b) they specify default scope when possible.{`} |
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3303/
PDF https://www.aclweb.org/anthology/W19-3303
PWC https://paperswithcode.com/paper/modeling-quantification-and-scope-in-abstract
Repo
Framework

Optimising the Machine Translation Post-editing Workflow

Title Optimising the Machine Translation Post-editing Workflow
Authors Anna Zaretskaya
Abstract In this article, we describe how machine translation is used for post-editing at TransPerfect and the ways in which we optimise the workflow. This includes MT evaluation, MT engine customisation, leveraging MT suggestions compared to TM matches, and the lessons learnt from implementing MT at a large scale.
Tasks Machine Translation
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-8718/
PDF https://www.aclweb.org/anthology/W19-8718
PWC https://paperswithcode.com/paper/optimising-the-machine-translation-post
Repo
Framework
comments powered by Disqus