October 16, 2019

2378 words 12 mins read

Paper Group NANR 6

Paper Group NANR 6

Attentional ShapeContextNet for Point Cloud Recognition. Learning a Warping Distance from Unlabeled Time Series Using Sequence Autoencoders. Proceedings of the First Workshop on Trolling, Aggression and Cyberbullying (TRAC-2018). CNN for Text-Based Multiple Choice Question Answering. Free Supervision From Video Games. A closer look at the word anal …

Attentional ShapeContextNet for Point Cloud Recognition

Title Attentional ShapeContextNet for Point Cloud Recognition
Authors Saining Xie, Sainan Liu, Zeyu Chen, Zhuowen Tu
Abstract We tackle the problem of point cloud recognition. Unlike previous approaches where a point cloud is either converted into a volume/image or represented independently in a permutation-invariant set, we develop a new representation by adopting the concept of shape context as the building block in our network design. The resulting model, called ShapeContextNet, consists of a hierarchy with modules not relying on a fixed grid while still enjoying properties similar to those in convolutional neural networks — being able to capture and propagate the object part information. In addition, we find inspiration from self-attention based models to include a simple yet effective contextual modeling mechanism — making the contextual region selection, the feature aggregation, and the feature transformation process fully automatic. ShapeContextNet is an end-to-end model that can be applied to the general point cloud classification and segmentation problems. We observe competitive results on a number of benchmark datasets.
Tasks
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/Xie_Attentional_ShapeContextNet_for_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/Xie_Attentional_ShapeContextNet_for_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/attentional-shapecontextnet-for-point-cloud
Repo
Framework

Learning a Warping Distance from Unlabeled Time Series Using Sequence Autoencoders

Title Learning a Warping Distance from Unlabeled Time Series Using Sequence Autoencoders
Authors Abubakar Abid, James Y. Zou
Abstract Measuring similarities between unlabeled time series trajectories is an important problem in many domains such as medicine, economics, and vision. It is often unclear what is the appropriate metric to use because of the complex nature of noise in the trajectories (e.g. different sampling rates or outliers). Experts typically hand-craft or manually select a specific metric, such as Dynamic Time Warping (DTW), to apply on their data. In this paper, we propose an end-to-end framework, autowarp, that optimizes and learns a good metric given unlabeled trajectories. We define a flexible and differentiable family of warping metrics, which encompasses common metrics such as DTW, Edit Distance, Euclidean, etc. Autowarp then leverages the representation power of sequence autoencoders to optimize for a member of this warping family. The output is an metric which is easy to interpret and can be robustly learned from relatively few trajectories. In systematic experiments across different domains, we show that autowarp often outperforms hand-crafted trajectory similarity metrics.
Tasks Time Series
Published 2018-12-01
URL http://papers.nips.cc/paper/8254-learning-a-warping-distance-from-unlabeled-time-series-using-sequence-autoencoders
PDF http://papers.nips.cc/paper/8254-learning-a-warping-distance-from-unlabeled-time-series-using-sequence-autoencoders.pdf
PWC https://paperswithcode.com/paper/learning-a-warping-distance-from-unlabeled
Repo
Framework

Proceedings of the First Workshop on Trolling, Aggression and Cyberbullying (TRAC-2018)

Title Proceedings of the First Workshop on Trolling, Aggression and Cyberbullying (TRAC-2018)
Authors
Abstract
Tasks
Published 2018-08-01
URL https://www.aclweb.org/anthology/W18-4400/
PDF https://www.aclweb.org/anthology/W18-4400
PWC https://paperswithcode.com/paper/proceedings-of-the-first-workshop-on-trolling
Repo
Framework

CNN for Text-Based Multiple Choice Question Answering

Title CNN for Text-Based Multiple Choice Question Answering
Authors Akshay Chaturvedi, P, Onkar it, Utpal Garain
Abstract The task of Question Answering is at the very core of machine comprehension. In this paper, we propose a Convolutional Neural Network (CNN) model for text-based multiple choice question answering where questions are based on a particular article. Given an article and a multiple choice question, our model assigns a score to each question-option tuple and chooses the final option accordingly. We test our model on Textbook Question Answering (TQA) and SciQ dataset. Our model outperforms several LSTM-based baseline models on the two datasets.
Tasks Question Answering, Reading Comprehension, Sentiment Analysis
Published 2018-07-01
URL https://www.aclweb.org/anthology/P18-2044/
PDF https://www.aclweb.org/anthology/P18-2044
PWC https://paperswithcode.com/paper/cnn-for-text-based-multiple-choice-question
Repo
Framework

Free Supervision From Video Games

Title Free Supervision From Video Games
Authors Philipp Krähenbühl
Abstract Deep networks are extremely hungry for data. They devour hundreds of thousands of labeled images to learn robust and semantically meaningful feature representations. Current networks are so data hungry that collecting labeled data has become as important as designing the networks themselves. Unfortunately, manual data collection is both expensive and time consuming. We present an alternative, and show how ground truth labels for many vision tasks are easily extracted from video games in real time as we play them. We interface the popular Microsoft DirectX rendering API, and inject specialized rendering code into the game as it is running. This code produces ground truth labels for instance segmentation, semantic labeling, depth estimation, optical flow, intrinsic image decomposition, and instance tracking. Instead of labeling images, a researcher now simply plays video games all day long. Our method is general and works on a wide range of video games. We collected a dataset of 220k training images, and 60k test images across 3 video games, and evaluate state of the art optical flow, depth estimation and intrinsic image decomposition algorithms. Our video game data is visually closer to real world images, than other synthetic dataset.
Tasks Depth Estimation, Instance Segmentation, Intrinsic Image Decomposition, Optical Flow Estimation, Semantic Segmentation
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/Krahenbuhl_Free_Supervision_From_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/Krahenbuhl_Free_Supervision_From_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/free-supervision-from-video-games
Repo
Framework

A closer look at the word analogy problem

Title A closer look at the word analogy problem
Authors Siddharth Krishna Kumar
Abstract Although word analogy problems have become a standard tool for evaluating word vectors, little is known about why word vectors are so good at solving these problems. In this paper, I attempt to further our understanding of the subject, by developing a simple, but highly accurate generative approach to solve the word analogy problem for the case when all terms involved in the problem are nouns. My results demonstrate the ambiguities associated with learning the relationship between a word pair, and the role of the training dataset in determining the relationship which gets most highlighted. Furthermore, my results show that the ability of a model to accurately solve the word analogy problem may not be indicative of a model’s ability to learn the relationship between a word pair the way a human does.
Tasks
Published 2018-01-01
URL https://openreview.net/forum?id=ryA-jdlA-
PDF https://openreview.net/pdf?id=ryA-jdlA-
PWC https://paperswithcode.com/paper/a-closer-look-at-the-word-analogy-problem
Repo
Framework

JeSemE: Interleaving Semantics and Emotions in a Web Service for the Exploration of Language Change Phenomena

Title JeSemE: Interleaving Semantics and Emotions in a Web Service for the Exploration of Language Change Phenomena
Authors Johannes Hellrich, Sven Buechel, Udo Hahn
Abstract We here introduce a substantially extended version of JeSemE, an interactive website for visually exploring computationally derived time-variant information on word meanings and lexical emotions assembled from five large diachronic text corpora. JeSemE is designed for scholars in the (digital) humanities as an alternative to consulting manually compiled, printed dictionaries for such information (if available at all). This tool uniquely combines state-of-the-art distributional semantics with a nuanced model of human emotions, two information streams we deem beneficial for a data-driven interpretation of texts in the humanities.
Tasks Sentiment Analysis, Word Embeddings
Published 2018-08-01
URL https://www.aclweb.org/anthology/C18-2003/
PDF https://www.aclweb.org/anthology/C18-2003
PWC https://paperswithcode.com/paper/jeseme-interleaving-semantics-and-emotions-in
Repo
Framework

Using Morphological Knowledge in Open-Vocabulary Neural Language Models

Title Using Morphological Knowledge in Open-Vocabulary Neural Language Models
Authors Austin Matthews, Graham Neubig, Chris Dyer
Abstract Languages with productive morphology pose problems for language models that generate words from a fixed vocabulary. Although character-based models allow any possible word type to be generated, they are linguistically na{"\i}ve: they must discover that words exist and are delimited by spaces{—}basic linguistic facts that are built in to the structure of word-based models. We introduce an open-vocabulary language model that incorporates more sophisticated linguistic knowledge by predicting words using a mixture of three generative processes: (1) by generating words as a sequence of characters, (2) by directly generating full word forms, and (3) by generating words as a sequence of morphemes that are combined using a hand-written morphological analyzer. Experiments on Finnish, Turkish, and Russian show that our model outperforms character sequence models and other strong baselines on intrinsic and extrinsic measures. Furthermore, we show that our model learns to exploit morphological knowledge encoded in the analyzer, and, as a byproduct, it can perform effective unsupervised morphological disambiguation.
Tasks Language Modelling
Published 2018-06-01
URL https://www.aclweb.org/anthology/N18-1130/
PDF https://www.aclweb.org/anthology/N18-1130
PWC https://paperswithcode.com/paper/using-morphological-knowledge-in-open
Repo
Framework

Efficient Gradient Computation for Structured Output Learning with Rational and Tropical Losses

Title Efficient Gradient Computation for Structured Output Learning with Rational and Tropical Losses
Authors Corinna Cortes, Vitaly Kuznetsov, Mehryar Mohri, Dmitry Storcheus, Scott Yang
Abstract Many structured prediction problems admit a natural loss function for evaluation such as the edit-distance or $n$-gram loss. However, existing learning algorithms are typically designed to optimize alternative objectives such as the cross-entropy. This is because a na"{i}ve implementation of the natural loss functions often results in intractable gradient computations. In this paper, we design efficient gradient computation algorithms for two broad families of structured prediction loss functions: rational and tropical losses. These families include as special cases the $n$-gram loss, the edit-distance loss, and many other loss functions commonly used in natural language processing and computational biology tasks that are based on sequence similarity measures. Our algorithms make use of weighted automata and graph operations over appropriate semirings to design efficient solutions. They facilitate efficient gradient computation and hence enable one to train learning models such as neural networks with complex structured losses.
Tasks Structured Prediction
Published 2018-12-01
URL http://papers.nips.cc/paper/7914-efficient-gradient-computation-for-structured-output-learning-with-rational-and-tropical-losses
PDF http://papers.nips.cc/paper/7914-efficient-gradient-computation-for-structured-output-learning-with-rational-and-tropical-losses.pdf
PWC https://paperswithcode.com/paper/efficient-gradient-computation-for-structured
Repo
Framework

NTT’s Neural Machine Translation Systems for WMT 2018

Title NTT’s Neural Machine Translation Systems for WMT 2018
Authors Makoto Morishita, Jun Suzuki, Masaaki Nagata
Abstract This paper describes NTT{'}s neural machine translation systems submitted to the WMT 2018 English-German and German-English news translation tasks. Our submission has three main components: the Transformer model, corpus cleaning, and right-to-left n-best re-ranking techniques. Through our experiments, we identified two keys for improving accuracy: filtering noisy training sentences and right-to-left re-ranking. We also found that the Transformer model requires more training data than the RNN-based model, and the RNN-based model sometimes achieves better accuracy than the Transformer model when the corpus is small.
Tasks Machine Translation
Published 2018-10-01
URL https://www.aclweb.org/anthology/W18-6421/
PDF https://www.aclweb.org/anthology/W18-6421
PWC https://paperswithcode.com/paper/ntts-neural-machine-translation-systems-for
Repo
Framework

Incorporating Semantic Attention in Video Description Generation

Title Incorporating Semantic Attention in Video Description Generation
Authors Natsuda Laokulrat, Naoaki Okazaki, Hideki Nakayama
Abstract
Tasks Image Captioning, Image Classification, Language Modelling, Object Recognition, Video Description
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1477/
PDF https://www.aclweb.org/anthology/L18-1477
PWC https://paperswithcode.com/paper/incorporating-semantic-attention-in-video
Repo
Framework

Geometry-Aware Scene Text Detection With Instance Transformation Network

Title Geometry-Aware Scene Text Detection With Instance Transformation Network
Authors Fangfang Wang, Liming Zhao, Xi Li, Xinchao Wang, Dacheng Tao
Abstract Localizing text in the wild is challenging in the situations of complicated geometric layout of the targets like random orientation and large aspect ratio. In this paper, we propose a geometry-aware modeling approach tailored for scene text representation with an end-to-end learning scheme. In our approach, a novel Instance Transformation Network (ITN) is presented to learn the geometry-aware representation encoding the unique geometric configurations of scene text instances with in-network transformation embedding, resulting in a robust and elegant framework to detect words or text lines at one pass. An end-to-end multi-task learning strategy with transformation regression, text/non-text classification and coordinate regression is adopted in the ITN. Experiments on the benchmark datasets demonstrate the effectiveness of the proposed approach in detecting scene text in various geometric configurations.
Tasks Multi-Task Learning, Scene Text Detection, Text Classification
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/Wang_Geometry-Aware_Scene_Text_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/Wang_Geometry-Aware_Scene_Text_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/geometry-aware-scene-text-detection-with
Repo
Framework

Clustering Redemption–Beyond the Impossibility of Kleinberg’s Axioms

Title Clustering Redemption–Beyond the Impossibility of Kleinberg’s Axioms
Authors Vincent Cohen-Addad, Varun Kanade, Frederik Mallmann-Trenn
Abstract Kleinberg (2002) stated three axioms that any clustering procedure should satisfy and showed there is no clustering procedure that simultaneously satisfies all three. One of these, called the consistency axiom, requires that when the data is modified in a helpful way, i.e. if points in the same cluster are made more similar and those in different ones made less similar, the algorithm should output the same clustering. To circumvent this impossibility result, research has focused on considering clustering procedures that have a clustering quality measure (or a cost) and showing that a modification of Kleinberg’s axioms that takes cost into account lead to feasible clustering procedures. In this work, we take a different approach, based on the observation that the consistency axiom fails to be satisfied when the “correct” number of clusters changes. We modify this axiom by making use of cost functions to determine the correct number of clusters, and require that consistency holds only if the number of clusters remains unchanged. We show that single linkage satisfies the modified axioms, and if the input is well-clusterable, some popular procedures such as k-means also satisfy the axioms, taking a step towards explaining the success of these objective functions for guiding the design of algorithms.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/8071-clustering-redemptionbeyond-the-impossibility-of-kleinbergs-axioms
PDF http://papers.nips.cc/paper/8071-clustering-redemptionbeyond-the-impossibility-of-kleinbergs-axioms.pdf
PWC https://paperswithcode.com/paper/clustering-redemptionbeyond-the-impossibility
Repo
Framework

AX Semantics’ Submission to the CoNLL–SIGMORPHON 2018 Shared Task

Title AX Semantics’ Submission to the CoNLL–SIGMORPHON 2018 Shared Task
Authors Andreas Madsack, Alessia Cavallo, Johanna Heininger, Robert Wei{\ss}graeber
Abstract
Tasks
Published 2018-10-01
URL https://www.aclweb.org/anthology/K18-3004/
PDF https://www.aclweb.org/anthology/K18-3004
PWC https://paperswithcode.com/paper/ax-semantics-submission-to-the-conll
Repo
Framework

What do I Annotate Next? An Empirical Study of Active Learning for Action Localization

Title What do I Annotate Next? An Empirical Study of Active Learning for Action Localization
Authors Fabian Caba Heilbron, Joon-Young Lee, Hailin Jin, Bernard Ghanem
Abstract Despite tremendous progress achieved in temporal action localization, state-of-the-art methods still struggle to train accurate models when annotated data is scarce. In this paper, we introduce a novel active learning framework for temporal localization that aims to mitigate this data dependency issue. We equip our framework with active selection functions that can reuse knowledge from previously annotated datasets. We study the performance of two state-of-the-art active selection functions as well as two widely used active learning baselines. To validate the effectiveness of each one of these selection functions, we conduct simulated experiments on ActivityNet. We find that using previously acquired knowledge as a bootstrapping source is crucial for active learners aiming to localize actions. When equipped with the right selection function, our proposed framework exhibits significantly better performance than standard active learning strategies, such as uncertainty sampling. Finally, we employ our framework to augment the newly compiled Kinetics action dataset with ground-truth temporal annotations. As a result, we collect Kinetics-Localization, a novel large-scale dataset for temporal action localization, which contains more than 15K YouTube videos.
Tasks Action Localization, Active Learning, Temporal Action Localization, Temporal Localization
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/Fabian_Caba_What_do_I_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/Fabian_Caba_What_do_I_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/what-do-i-annotate-next-an-empirical-study-of
Repo
Framework
comments powered by Disqus