January 25, 2020

2537 words 12 mins read

Paper Group NANR 77

Paper Group NANR 77

Localization of Deep Inpainting Using High-Pass Fully Convolutional Network. KU_ai at MEDIQA 2019: Domain-specific Pre-training and Transfer Learning for Medical NLI. Parallel Optimal Transport GAN. An Analysis of Composite Neural Network Performance from Function Composition Perspective. Cross-lingual Transfer Learning and Multitask Learning for …

Localization of Deep Inpainting Using High-Pass Fully Convolutional Network

Title Localization of Deep Inpainting Using High-Pass Fully Convolutional Network
Authors Haodong Li, Jiwu Huang
Abstract Image inpainting has been substantially improved with deep learning in the past years. Deep inpainting can fill image regions with plausible contents, which are not visually apparent. Although inpainting is originally designed to repair images, it can even be used for malicious manipulations, e.g., removal of specific objects. Therefore, it is necessary to identify the presence of inpainting in an image. This paper presents a method to locate the regions manipulated by deep inpainting. The proposed method employs a fully convolutional network that is based on high-pass filtered image residuals. Firstly, we analyze and observe that the inpainted regions are more distinguishable from the untouched ones in the residual domain. Hence, a high-pass pre-filtering module is designed to get image residuals for enhancing inpainting traces. Then, a feature extraction module, which learns discriminative features from image residuals, is built with four concatenated ResNet blocks. The learned feature maps are finally enlarged by an up-sampling module, so that a pixel-wise inpainting localization map is obtained. The whole network is trained end-to-end with a loss addressing the class imbalance. Extensive experimental results evaluated on both synthetic and realistic images subjected to deep inpainting have shown the effectiveness of the proposed method.
Tasks Image Inpainting
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Li_Localization_of_Deep_Inpainting_Using_High-Pass_Fully_Convolutional_Network_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Li_Localization_of_Deep_Inpainting_Using_High-Pass_Fully_Convolutional_Network_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/localization-of-deep-inpainting-using-high
Repo
Framework

KU_ai at MEDIQA 2019: Domain-specific Pre-training and Transfer Learning for Medical NLI

Title KU_ai at MEDIQA 2019: Domain-specific Pre-training and Transfer Learning for Medical NLI
Authors Cemil Cengiz, Ula{\c{s}} Sert, Deniz Yuret
Abstract In this paper, we describe our system and results submitted for the Natural Language Inference (NLI) track of the MEDIQA 2019 Shared Task. As KU{_}ai team, we used BERT as our baseline model and pre-processed the MedNLI dataset to mitigate the negative impact of de-identification artifacts. Moreover, we investigated different pre-training and transfer learning approaches to improve the performance. We show that pre-training the language model on rich biomedical corpora has a significant effect in teaching the model domain-specific language. In addition, training the model on large NLI datasets such as MultiNLI and SNLI helps in learning task-specific reasoning. Finally, we ensembled our highest-performing models, and achieved 84.7{%} accuracy on the unseen test dataset and ranked 10th out of 17 teams in the official results.
Tasks Language Modelling, Natural Language Inference, Transfer Learning
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5045/
PDF https://www.aclweb.org/anthology/W19-5045
PWC https://paperswithcode.com/paper/ku_ai-at-mediqa-2019-domain-specific-pre
Repo
Framework

Parallel Optimal Transport GAN

Title Parallel Optimal Transport GAN
Authors Gil Avraham, Yan Zuo, Tom Drummond
Abstract Although Generative Adversarial Networks (GANs) are known for their sharp realism in image generation, they often fail to estimate areas of the data density. This leads to low modal diversity and at times distorted generated samples. These problems essentially arise from poor estimation of the distance metric responsible for training these networks. To address these issues, we introduce an additional regularisation term which performs optimal transport in parallel within a low dimensional representation space. We demonstrate that operating in a low dimension representation of the data distribution benefits from convergence rate gains in estimating the Wasserstein distance, resulting in more stable GAN training. We empirically show that our regulariser achieves a stabilising effect which leads to higher quality of generated samples and increased mode coverage of the given data distribution. Our method achieves significant improvements on the CIFAR-10, Oxford Flowers and CUB Birds datasets over several GAN baselines both qualitatively and quantitatively.
Tasks Image Generation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Avraham_Parallel_Optimal_Transport_GAN_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Avraham_Parallel_Optimal_Transport_GAN_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/parallel-optimal-transport-gan
Repo
Framework

An Analysis of Composite Neural Network Performance from Function Composition Perspective

Title An Analysis of Composite Neural Network Performance from Function Composition Perspective
Authors Ming-Chuan Yang, Meng Chang Chen
Abstract This work investigates the performance of a composite neural network, which is composed of pre-trained neural network models and non-instantiated neural network models, connected to form a rooted directed graph. A pre-trained neural network model is generally a well trained neural network model targeted for a specific function. The advantages of adopting such a pre-trained model in a composite neural network are two folds. One is to benefit from other’s intelligence and diligence and the other is saving the efforts in data preparation and resources and time in training. However, the overall performance of composite neural network is still not clear. In this work, we prove that a composite neural network, with high probability, performs better than any of its pre-trained components under certain assumptions. In addition, if an extra pre-trained component is added to a composite network, with high probability the overall performance will be improved. In the empirical evaluations, distinctively different applications support the above findings.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=HkGSniC9FQ
PDF https://openreview.net/pdf?id=HkGSniC9FQ
PWC https://paperswithcode.com/paper/an-analysis-of-composite-neural-network
Repo
Framework

Cross-lingual Transfer Learning and Multitask Learning for Capturing Multiword Expressions

Title Cross-lingual Transfer Learning and Multitask Learning for Capturing Multiword Expressions
Authors Shiva Taslimipoor, Omid Rohanian, Le An Ha
Abstract Recent developments in deep learning have prompted a surge of interest in the application of multitask and transfer learning to NLP problems. In this study, we explore for the first time, the application of transfer learning (TRL) and multitask learning (MTL) to the identification of Multiword Expressions (MWEs). For MTL, we exploit the shared syntactic information between MWE and dependency parsing models to jointly train a single model on both tasks. We specifically predict two types of labels: MWE and dependency parse. Our neural MTL architecture utilises the supervision of dependency parsing in lower layers and predicts MWE tags in upper layers. In the TRL scenario, we overcome the scarcity of data by learning a model on a larger MWE dataset and transferring the knowledge to a resource-poor setting in another language. In both scenarios, the resulting models achieved higher performance compared to standard neural approaches.
Tasks Cross-Lingual Transfer, Dependency Parsing, Transfer Learning
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5119/
PDF https://www.aclweb.org/anthology/W19-5119
PWC https://paperswithcode.com/paper/cross-lingual-transfer-learning-and-multitask
Repo
Framework

Towards Assessing Argumentation Annotation - A First Step

Title Towards Assessing Argumentation Annotation - A First Step
Authors Anna Lindahl, Lars Borin, Jacobo Rouces
Abstract This paper presents a first attempt at using Walton{'}s argumentation schemes for annotating arguments in Swedish political text and assessing the feasibility of using this particular set of schemes with two linguistically trained annotators. The texts are not pre-annotated with argumentation structure beforehand. The results show that the annotators differ both in number of annotated arguments and selection of the conclusion and premises which make up the arguments. They also differ in their labeling of the schemes, but grouping the schemes increases their agreement. The outcome from this will be used to develop guidelines for future annotations.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4520/
PDF https://www.aclweb.org/anthology/W19-4520
PWC https://paperswithcode.com/paper/towards-assessing-argumentation-annotation-a
Repo
Framework

The Comparative Power of ReLU Networks and Polynomial Kernels in the Presence of Sparse Latent Structure

Title The Comparative Power of ReLU Networks and Polynomial Kernels in the Presence of Sparse Latent Structure
Authors Frederic Koehler, Andrej Risteski
Abstract There has been a large amount of interest, both in the past and particularly recently, into the relative advantage of different families of universal function approximators, for instance neural networks, polynomials, rational functions, etc. However, current research has focused almost exclusively on understanding this problem in a worst case setting: e.g. characterizing the best L1 or L_{infty} approximation in a box (or sometimes, even under an adversarially constructed data distribution.) In this setting many classical tools from approximation theory can be effectively used. However, in typical applications we expect data to be high dimensional, but structured – so, it would only be important to approximate the desired function well on the relevant part of its domain, e.g. a small manifold on which real input data actually lies. Moreover, even within this domain the desired quality of approximation may not be uniform; for instance in classification problems, the approximation needs to be more accurate near the decision boundary. These issues, to the best of our knowledge, have remain unexplored until now. With this in mind, we analyze the performance of neural networks and polynomial kernels in a natural regression setting where the data enjoys sparse latent structure, and the labels depend in a simple way on the latent variables. We give an almost-tight theoretical analysis of the performance of both neural networks and polynomials for this problem, as well as verify our theory with simulations. Our results both involve new (complex-analytic) techniques, which may be of independent interest, and show substantial qualitative differences with what is known in the worst-case setting.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=rJgTTjA9tX
PDF https://openreview.net/pdf?id=rJgTTjA9tX
PWC https://paperswithcode.com/paper/the-comparative-power-of-relu-networks-and
Repo
Framework

P=aṇinian Syntactico-Semantic Relation Labels

Title P=aṇinian Syntactico-Semantic Relation Labels
Authors Amba Kulkarni, Dipti Sharma
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-7724/
PDF https://www.aclweb.org/anthology/W19-7724
PWC https://paperswithcode.com/paper/paa1inian-syntactico-semantic-relation-labels
Repo
Framework

Graph convolutional networks for exploring authorship hypotheses

Title Graph convolutional networks for exploring authorship hypotheses
Authors Tom Lippincott
Abstract This work considers a task from traditional literary criticism: annotating a structured, composite document with information about its sources. We take the Documentary Hypothesis, a prominent theory regarding the composition of the first five books of the Hebrew bible, extract stylistic features designed to avoid bias or overfitting, and train several classification models. Our main result is that the recently-introduced graph convolutional network architecture outperforms structurally-uninformed models. We also find that including information about the granularity of text spans is a crucial ingredient when employing hidden layers, in contrast to simple logistic regression. We perform error analysis at several levels, noting how some characteristic limitations of the models and simple features lead to misclassifications, and conclude with an overview of future work.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-2510/
PDF https://www.aclweb.org/anthology/W19-2510
PWC https://paperswithcode.com/paper/graph-convolutional-networks-for-exploring
Repo
Framework

Asking the Crowd: Question Analysis, Evaluation and Generation for Open Discussion on Online Forums

Title Asking the Crowd: Question Analysis, Evaluation and Generation for Open Discussion on Online Forums
Authors Zi Chai, Xinyu Xing, Xiaojun Wan, Bo Huang
Abstract Teaching machines to ask questions is an important yet challenging task. Most prior work focused on generating questions with fixed answers. As contents are highly limited by given answers, these questions are often not worth discussing. In this paper, we take the first step on teaching machines to ask open-answered questions from real-world news for open discussion (openQG). To generate high-qualified questions, effective ways for question evaluation are required. We take the perspective that the more answers a question receives, the better it is for open discussion, and analyze how language use affects the number of answers. Compared with other factors, e.g. topic and post time, linguistic factors keep our evaluation from being domain-specific. We carefully perform variable control on 11.5M questions from online forums to get a dataset, OQRanD, and further perform question analysis. Based on these conclusions, several models are built for question evaluation. For openQG task, we construct OQGenD, the first dataset as far as we know, and propose a model based on conditional generative adversarial networks and our question evaluation model. Experiments show that our model can generate questions with higher quality compared with commonly-used text generation methods.
Tasks Text Generation
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1497/
PDF https://www.aclweb.org/anthology/P19-1497
PWC https://paperswithcode.com/paper/asking-the-crowd-question-analysis-evaluation
Repo
Framework

Decision-making, Risk, and Gist Machine Translation in the Work of Patent Professionals

Title Decision-making, Risk, and Gist Machine Translation in the Work of Patent Professionals
Authors Mary Nurminen
Abstract
Tasks Decision Making, Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-7204/
PDF https://www.aclweb.org/anthology/W19-7204
PWC https://paperswithcode.com/paper/decision-making-risk-and-gist-machine
Repo
Framework

Constrained Sequence-to-sequence Semitic Root Extraction for Enriching Word Embeddings

Title Constrained Sequence-to-sequence Semitic Root Extraction for Enriching Word Embeddings
Authors Ahmed El-Kishky, Xingyu Fu, Aseel Addawood, Nahil Sobh, Clare Voss, Jiawei Han
Abstract In this paper, we tackle the problem of {``}root extraction{''} from words in the Semitic language family. A challenge in applying natural language processing techniques to these languages is the data sparsity problem that arises from their rich internal morphology, where the substructure is inherently non-concatenative and morphemes are interdigitated in word formation. While previous automated methods have relied on human-curated rules or multiclass classification, they have not fully leveraged the various combinations of regular, sequential concatenative morphology within the words and the internal interleaving within templatic stems of roots and patterns. To address this, we propose a constrained sequence-to-sequence root extraction method. Experimental results show our constrained model outperforms a variety of methods at root extraction. Furthermore, by enriching word embeddings with resulting decompositions, we show improved results on word analogy, word similarity, and language modeling tasks. |
Tasks Language Modelling, Word Embeddings
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4610/
PDF https://www.aclweb.org/anthology/W19-4610
PWC https://paperswithcode.com/paper/constrained-sequence-to-sequence-semitic-root
Repo
Framework

Proceedings of the 1st International Workshop of AI Werewolf and Dialog System (AIWolfDial2019)

Title Proceedings of the 1st International Workshop of AI Werewolf and Dialog System (AIWolfDial2019)
Authors
Abstract
Tasks
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-8300/
PDF https://www.aclweb.org/anthology/W19-8300
PWC https://paperswithcode.com/paper/proceedings-of-the-1st-international-workshop-1
Repo
Framework

An automatic discourse relation alignment experiment on TED-MDB

Title An automatic discourse relation alignment experiment on TED-MDB
Authors Sibel Ozer, Deniz Zeyrek
Abstract This paper describes an automatic discourse relation alignment experiment as an empirical justification of the planned annotation projection approach to enlarge the 3600-word multilingual corpus of TED Multilingual Discourse Bank (TED-MDB). The experiment is carried out on a single language pair (English-Turkish) included in TED-MDB. The paper first describes the creation of a large corpus of English-Turkish bi-sentences, then it presents a sense-based experiment that automatically aligns the relations in the English sentences of TED-MDB with the Turkish sentences. The results are very close to the results obtained from an earlier semi-automatic post-annotation alignment experiment validated by human annotators and are encouraging for future annotation projection tasks.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/papers/W/W19/W19-3612/
PDF https://www.aclweb.org/anthology/W19-3612
PWC https://paperswithcode.com/paper/an-automatic-discourse-relation-alignment
Repo
Framework

Formality Style Transfer for Noisy, User-generated Conversations: Extracting Labeled, Parallel Data from Unlabeled Corpora

Title Formality Style Transfer for Noisy, User-generated Conversations: Extracting Labeled, Parallel Data from Unlabeled Corpora
Authors Isak Czeresnia Etinger, Alan W Black
Abstract Typical datasets used for style transfer in NLP contain aligned pairs of two opposite extremes of a style. As each existing dataset is sourced from a specific domain and context, most use cases will have a sizable mismatch from the vocabulary and sentence structures of any dataset available. This reduces the performance of the style transfer, and is particularly significant for noisy, user-generated text. To solve this problem, we show a technique to derive a dataset of aligned pairs (style-agnostic vs stylistic sentences) from an unlabeled corpus by using an auxiliary dataset, allowing for in-domain training. We test the technique with the Yahoo Formality Dataset and 6 novel datasets we produced, which consist of scripts from 5 popular TV-shows (Friends, Futurama, Seinfeld, Southpark, Stargate SG-1) and the Slate Star Codex online forum. We gather 1080 human evaluations, which show that our method produces a sizable change in formality while maintaining fluency and context; and that it considerably outperforms OpenNMT{'}s Seq2Seq model directly trained on the Yahoo Formality Dataset. Additionally, we publish the full pipeline code and our novel datasets.
Tasks Style Transfer
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5502/
PDF https://www.aclweb.org/anthology/D19-5502
PWC https://paperswithcode.com/paper/formality-style-transfer-for-noisy-user
Repo
Framework
comments powered by Disqus