October 15, 2019

2470 words 12 mins read

Paper Group NANR 93

Paper Group NANR 93

Transfer Learning on Manifolds via Learned Transport Operators. The Information-Autoencoding Family: A Lagrangian Perspective on Latent Variable Generative Modeling. Training deep learning based denoisers without ground truth data. Stacked Latent Attention for Multimodal Reasoning. Causality Analysis of Twitter Sentiments and Stock Market Returns. …

Transfer Learning on Manifolds via Learned Transport Operators

Title Transfer Learning on Manifolds via Learned Transport Operators
Authors Marissa Connor, Christopher Rozell
Abstract Within-class variation in a high-dimensional dataset can be modeled as being on a low-dimensional manifold due to the constraints of the physical processes producing that variation (e.g., translation, illumination, etc.). We desire a method for learning a representation of the manifolds induced by identity-preserving transformations that can be used to increase robustness, reduce the training burden, and encourage interpretability in machine learning tasks. In particular, what is needed is a representation of the transformation manifold that can robustly capture the shape of the manifold from the input data, generate new points on the manifold, and extend transformations outside of the training domain without significantly increasing the error. Previous work has proposed algorithms to efficiently learn analytic operators (called transport operators) that define the process of transporting one data point on a manifold to another. The main contribution of this paper is to define two transfer learning methods that use this generative manifold representation to learn natural transformations and incorporate them into new data. The first method uses this representation in a novel randomized approach to transfer learning that employs the learned generative model to map out unseen regions of the data space. These results are shown through demonstrations of transfer learning in a data augmentation task for few-shot image classification. The second method use of transport operators for injecting specific transformations into new data examples which allows for realistic image animation and informed data augmentation. These results are shown on stylized constructions using the classic swiss roll data structure and in demonstrations of transfer learning in a data augmentation task for few-shot image classification. We also propose the use of transport operators for injecting transformations into new data examples which allows for realistic image animation.
Tasks Data Augmentation, Few-Shot Image Classification, Image Animation, Image Classification, Transfer Learning
Published 2018-01-01
URL https://openreview.net/forum?id=rJL6pz-CZ
PDF https://openreview.net/pdf?id=rJL6pz-CZ
PWC https://paperswithcode.com/paper/transfer-learning-on-manifolds-via-learned
Repo
Framework

The Information-Autoencoding Family: A Lagrangian Perspective on Latent Variable Generative Modeling

Title The Information-Autoencoding Family: A Lagrangian Perspective on Latent Variable Generative Modeling
Authors Shengjia Zhao, Jiaming Song, Stefano Ermon
Abstract A variety of learning objectives have been recently proposed for training generative models. We show that many of them, including InfoGAN, ALI/BiGAN, ALICE, CycleGAN, VAE, $\beta$-VAE, adversarial autoencoders, AVB, and InfoVAE, are Lagrangian duals of the same primal optimization problem. This generalization reveals the implicit modeling trade-offs between flexibility and computational requirements being made by these models. Furthermore, we characterize the class of all objectives that can be optimized under certain computational constraints. Finally, we show how this new Lagrangian perspective can explain undesirable behavior of existing methods and provide new principled solutions.
Tasks
Published 2018-01-01
URL https://openreview.net/forum?id=ryZERzWCZ
PDF https://openreview.net/pdf?id=ryZERzWCZ
PWC https://paperswithcode.com/paper/the-information-autoencoding-family-a-1
Repo
Framework

Training deep learning based denoisers without ground truth data

Title Training deep learning based denoisers without ground truth data
Authors Shakarim Soltanayev, Se Young Chun
Abstract Recently developed deep-learning-based denoisers often outperform state-of-the-art conventional denoisers, such as the BM3D. They are typically trained to minimizethe mean squared error (MSE) between the output image of a deep neural networkand a ground truth image. In deep learning based denoisers, it is important to use high quality noiseless ground truth data for high performance, but it is often challenging or even infeasible to obtain noiseless images in application areas such as hyperspectral remote sensing and medical imaging. In this article, we propose a method based on Stein’s unbiased risk estimator (SURE) for training deep neural network denoisers only based on the use of noisy images. We demonstrate that our SURE-based method, without the use of ground truth data, is able to train deep neural network denoisers to yield performances close to those networks trained with ground truth, and to outperform the state-of-the-art denoiser BM3D. Further improvements were achieved when noisy test images were used for training of denoiser networks using our proposed SURE-based method.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/7587-training-deep-learning-based-denoisers-without-ground-truth-data
PDF http://papers.nips.cc/paper/7587-training-deep-learning-based-denoisers-without-ground-truth-data.pdf
PWC https://paperswithcode.com/paper/training-deep-learning-based-denoisers-1
Repo
Framework

Stacked Latent Attention for Multimodal Reasoning

Title Stacked Latent Attention for Multimodal Reasoning
Authors Haoqi Fan, Jiatong Zhou
Abstract Attention has shown to be a pivotal development in deep learning and has been used for a multitude of multimodal learning tasks such as visual question answering and image captioning. In this work, we pinpoint the potential limitations to the design of a traditional attention model. We identify that 1) current attention mechanisms discard the latent information from intermediate reasoning, losing the positional information already captured by the attention heatmaps and 2) stacked attention, a common way to improve spatial reasoning, may have suboptimal performance because of the vanishing gradient problem. We introduce a novel attention architecture to address these problems, in which all spatial configuration information contained in the intermediate reasoning process is retained in a pathway of convolutional layers. We show that this new attention leads to substantial improvements in multiple multimodal reasoning tasks, including achieving single model performance without using external knowledge comparable to the state-of-the-art on the VQA dataset, as well as clear gains for the image captioning task.
Tasks Image Captioning, Question Answering, Visual Question Answering
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/Fan_Stacked_Latent_Attention_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/Fan_Stacked_Latent_Attention_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/stacked-latent-attention-for-multimodal
Repo
Framework

Causality Analysis of Twitter Sentiments and Stock Market Returns

Title Causality Analysis of Twitter Sentiments and Stock Market Returns
Authors Narges Tabari, Piyusha Biswas, Bhanu Praneeth, Armin Seyeditabari, Mirsad Hadzikadic, Wlodek Zadrozny
Abstract Sentiment analysis is the process of identifying the opinion expressed in text. Recently, it has been used to study behavioral finance, and in particular the effect of opinions and emotions on economic or financial decisions. In this paper, we use a public dataset of labeled tweets that has been labeled by Amazon Mechanical Turk and then we propose a baseline classification model. Then, by using Granger causality of both sentiment datasets with the different stocks, we shows that there is causality between social media and stock market returns (in both directions) for many stocks. Finally, We evaluate this causality analysis by showing that in the event of a specific news on certain dates, there are evidences of trending the same news on Twitter for that stock.
Tasks Sentiment Analysis, Twitter Sentiment Analysis
Published 2018-07-01
URL https://www.aclweb.org/anthology/W18-3102/
PDF https://www.aclweb.org/anthology/W18-3102
PWC https://paperswithcode.com/paper/causality-analysis-of-twitter-sentiments-and
Repo
Framework

When ACE met KBP: End-to-End Evaluation of Knowledge Base Population with Component-level Annotation

Title When ACE met KBP: End-to-End Evaluation of Knowledge Base Population with Component-level Annotation
Authors Bonan Min, Marjorie Freedman, Roger Bock, Ralph Weischedel
Abstract
Tasks Knowledge Base Population, Named Entity Recognition, Question Answering, Relation Extraction, Slot Filling
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1031/
PDF https://www.aclweb.org/anthology/L18-1031
PWC https://paperswithcode.com/paper/when-ace-met-kbp-end-to-end-evaluation-of
Repo
Framework

Toward An Epic Epigraph Graph

Title Toward An Epic Epigraph Graph
Authors Francis Bond, Graham Matthews
Abstract
Tasks
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1522/
PDF https://www.aclweb.org/anthology/L18-1522
PWC https://paperswithcode.com/paper/toward-an-epic-epigraph-graph
Repo
Framework

TakeLab at SemEval-2018 Task12: Argument Reasoning Comprehension with Skip-Thought Vectors

Title TakeLab at SemEval-2018 Task12: Argument Reasoning Comprehension with Skip-Thought Vectors
Authors Ana Brassard, Tin Kuculo, Filip Boltu{\v{z}}i{'c}, Jan {\v{S}}najder
Abstract This paper describes our system for the SemEval-2018 Task 12: Argument Reasoning Comprehension Task. We utilize skip-thought vectors, sentence-level distributional vectors inspired by the popular word embeddings and the skip-gram model. We encode preprocessed sentences from the dataset into vectors, then perform a binary supervised classification of the warrant that justifies the use of the reason as support for the claim. We explore a few variations of the model, reaching 54.1{%} accuracy on the test set, which placed us 16th out of 22 teams participating in the task.
Tasks Common Sense Reasoning, Natural Language Inference, Word Embeddings
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-1192/
PDF https://www.aclweb.org/anthology/S18-1192
PWC https://paperswithcode.com/paper/takelab-at-semeval-2018-task12-argument
Repo
Framework

Universal Language Model Fine-tuning for Patent Classification

Title Universal Language Model Fine-tuning for Patent Classification
Authors Jason Hepburn
Abstract This paper describes the methods used for the 2018 ALTA Shared Task. The task this year was to automatically classify Australian patents into their main International Patent Classification section. Our final submission used a Support Vector Machine (SVM) and Universal Language Model with Fine-tuning (ULMFiT). Our system achieved the best results in the student category.
Tasks Language Modelling
Published 2018-12-01
URL https://www.aclweb.org/anthology/U18-1013/
PDF https://www.aclweb.org/anthology/U18-1013
PWC https://paperswithcode.com/paper/universal-language-model-fine-tuning-for
Repo
Framework

Generative Semantic Manipulation with Mask-Contrasting GAN

Title Generative Semantic Manipulation with Mask-Contrasting GAN
Authors Xiaodan Liang, Hao Zhang, Liang Lin, Eric Xing
Abstract Despite the promising results on paired/unpaired image-to-image translation achieved by Generative Adversarial Networks (GANs), prior works often only transfer the low-level information (e.g. color or texture changes), but fail to manipulate high-level semantic meanings (e.g., geometric structure or content) of different object regions. On the other hand, while some researches can synthesize compelling real-world images given a class label or caption, they cannot condition on arbitrary shapes or structures, which largely limits their application scenarios and interpretive capability of model results. In this work, we focus on a more challenging semantic manipulation task, aiming at modifying the semantic meaning of an object while preserving its own characteristics (e.g. viewpoints and shapes), such as cow$ ightarrow$sheep, motor$ ightarrow$ bicycle, cat$ ightarrow$dog. To tackle such large semantic changes, we introduce a contrasting GAN (contrast-GAN) with a novel adversarial contrasting objective which is able to perform all types of semantic translations with one category-conditional generator. Instead of directly making the synthesized samples close to target data as previous GANs did, our adversarial contrasting objective optimizes over the distance comparisons between samples, that is, enforcing the manipulated data be semantically closer to the real data with target category than the input data. Equipped with the new contrasting objective, a novel mask-conditional contrast-GAN architecture is proposed to enable disentangle image background with object semantic changes. Extensive qualitative and quantitative experiments on several semantic manipulation tasks on ImageNet and MSCOCO dataset show considerable performance gain by our contrast-GAN over other conditional GANs.
Tasks Image-to-Image Translation
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/Liang_Generative_Semantic_Manipulation_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/Liang_Generative_Semantic_Manipulation_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/generative-semantic-manipulation-with-mask
Repo
Framework

Multiple-Step Greedy Policies in Approximate and Online Reinforcement Learning

Title Multiple-Step Greedy Policies in Approximate and Online Reinforcement Learning
Authors Yonathan Efroni, Gal Dalal, Bruno Scherrer, Shie Mannor
Abstract Multiple-step lookahead policies have demonstrated high empirical competence in Reinforcement Learning, via the use of Monte Carlo Tree Search or Model Predictive Control. In a recent work (Efroni et al., 2018), multiple-step greedy policies and their use in vanilla Policy Iteration algorithms were proposed and analyzed. In this work, we study multiple-step greedy algorithms in more practical setups. We begin by highlighting a counter-intuitive difficulty, arising with soft-policy updates: even in the absence of approximations, and contrary to the 1-step-greedy case, monotonic policy improvement is not guaranteed unless the update stepsize is sufficiently large. Taking particular care about this difficulty, we formulate and analyze online and approximate algorithms that use such a multi-step greedy operator.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/7770-multiple-step-greedy-policies-in-approximate-and-online-reinforcement-learning
PDF http://papers.nips.cc/paper/7770-multiple-step-greedy-policies-in-approximate-and-online-reinforcement-learning.pdf
PWC https://paperswithcode.com/paper/multiple-step-greedy-policies-in-approximate
Repo
Framework

A Comparison of Entity Matching Methods between English and Japanese Katakana

Title A Comparison of Entity Matching Methods between English and Japanese Katakana
Authors Michiharu Yamashita, Hideki Awashima, Hidekazu Oiwa
Abstract Japanese Katakana is one component of the Japanese writing system and is used to express English terms, loanwords, and onomatopoeia in Japanese characters based on the phonemes. The main purpose of this research is to find the best entity matching methods between English and Katakana. We built two research questions to clarify which types of entity matching systems works better than others. The first question is what transliteration should be used for conversion. We need to transliterate English or Katakana terms into the same form in order to compute the string similarity. We consider five conversions that transliterate English to Katakana directly, Katakana to English directly, English to Katakana via phoneme, Katakana to English via phoneme, and both English and Katakana to phoneme. The second question is what should be used for the similarity measure at entity matching. To investigate the problem, we choose six methods, which are Overlap Coefficient, Cosine, Jaccard, Jaro-Winkler, Levenshtein, and the similarity of the phoneme probability predicted by RNN. Our results show that 1) matching using phonemes and conversion of Katakana to English works better than other methods, and 2) the similarity of phonemes outperforms other methods while other similarity score is changed depending on data and models.
Tasks Transliteration
Published 2018-10-01
URL https://www.aclweb.org/anthology/W18-5809/
PDF https://www.aclweb.org/anthology/W18-5809
PWC https://paperswithcode.com/paper/a-comparison-of-entity-matching-methods
Repo
Framework

The Circumstantial Event Ontology (CEO) and ECB+/CEO: an Ontology and Corpus for Implicit Causal Relations between Events

Title The Circumstantial Event Ontology (CEO) and ECB+/CEO: an Ontology and Corpus for Implicit Causal Relations between Events
Authors Roxane Segers, Tommaso Caselli, Piek Vossen
Abstract
Tasks Semantic Role Labeling
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1725/
PDF https://www.aclweb.org/anthology/L18-1725
PWC https://paperswithcode.com/paper/the-circumstantial-event-ontology-ceo-and
Repo
Framework

Spotting Spurious Data with Neural Networks

Title Spotting Spurious Data with Neural Networks
Authors Hadi Amiri, Timothy Miller, Guergana Savova
Abstract Automatic identification of spurious instances (those with potentially wrong labels in datasets) can improve the quality of existing language resources, especially when annotations are obtained through crowdsourcing or automatically generated based on coded rankings. In this paper, we present effective approaches inspired by queueing theory and psychology of learning to automatically identify spurious instances in datasets. Our approaches discriminate instances based on their {``}difficulty to learn,{''} determined by a downstream learner. Our methods can be applied to any dataset assuming the existence of a neural network model for the target task of the dataset. Our best approach outperforms competing state-of-the-art baselines and has a MAP of 0.85 and 0.22 in identifying spurious instances in synthetic and carefully-crowdsourced real-world datasets respectively. |
Tasks Object Classification
Published 2018-06-01
URL https://www.aclweb.org/anthology/N18-1182/
PDF https://www.aclweb.org/anthology/N18-1182
PWC https://paperswithcode.com/paper/spotting-spurious-data-with-neural-networks
Repo
Framework

Efficient 6-DoF Tracking of Handheld Objects from an Egocentric Viewpoint

Title Efficient 6-DoF Tracking of Handheld Objects from an Egocentric Viewpoint
Authors Rohit Pandey, Pavel Pidlypenskyi, Shuoran Yang, Christine Kaeser-Chen
Abstract Virtual and augmented reality technologies have seen significant growth in the past few years. A key component of such systems is the ability to track the pose of head mounted displays and controllers in 3D space. We tackle the problem of efficient 6-DoF tracking of a handheld controller from egocentric camera perspectives. We collected the HMD Controller dataset which consist of over 540,000 stereo image pairs labelled with the full 6-DoF pose of the handheld controller. Our proposed SSD-AF-Stereo3D model achieves a mean average error of 33.5 millimeters in 3D keypoint prediction and is used in conjunction with an IMU sensor on the controller to enable 6-DoF tracking. We also present results on approaches for model based full 6-DoF tracking. All our models operate under the strict constraints of real time mobile CPU inference.
Tasks
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/Rohit_Pandey_Efficient_6-DoF_Tracking_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/Rohit_Pandey_Efficient_6-DoF_Tracking_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/efficient-6-dof-tracking-of-handheld-objects
Repo
Framework
comments powered by Disqus