July 26, 2019

2715 words 13 mins read

Paper Group NAWR 1

Paper Group NAWR 1

Learning a model of facial shape and expression from 4D scans. AutoLearn - Automated Feature Generation and Selection. DataStories at SemEval-2017 Task 4: Deep LSTM with Attention for Message-level and Topic-based Sentiment Analysis. Grammatical Error Detection Using Error- and Grammaticality-Specific Word Embeddings. How Many Stemmata with Root De …

Learning a model of facial shape and expression from 4D scans

Title Learning a model of facial shape and expression from 4D scans
Authors Tianye Li, Timo Bolkart, Michael J. Black, Hao Li, Javier Romero
Abstract he field of 3D face modeling has a large gap between high-end and low-end methods. At the high end, the best facial animation is indistinguishablefrom real humans, but this comes at the cost of extensive manual labor. At the low end, face capture from consumer depth sensors relies on 3D face models that are not expressive enough to capture the variability in natural facial shape and expression. We seek a middle ground by learning a facial model from thousands of accurately aligned 3D scans. Our FLAME model (Faces Learned with an Articulated Model and Expressions) is designed to work with existing graphics software and be easy to fit to data. FLAME uses a linear shape space trained from 3800 scans of human heads. FLAME combines this linear shape space with an articulated jaw, neck, and eyeballs, pose-dependent corrective blendshapes, and additional global expression blendshapes. The pose and expression dependent articulations are learned from 4D face sequences in the D3DFACS dataset along with additional 4D sequences. We accurately register a template mesh to the scan sequences and make the D3DFACS registrations available for research purposes. In total the model is trained from over 33, 000 scans. FLAME is low-dimensional but more expressive than the FaceWarehouse model and the Basel Face Model. We compare FLAME to these models by fitting them to static 3D scans and 4D sequences using the same optimization method. FLAME is significantly more accurate and is available for research purposes.
Tasks 3D Face Animation, 3D Face Reconstruction, Face Generation
Published 2017-11-01
URL http://flame.is.tue.mpg.de/
PDF https://ps.is.tuebingen.mpg.de/uploads_file/attachment/attachment/400/paper.pdf
PWC https://paperswithcode.com/paper/learning-a-model-of-facial-shape-and
Repo https://github.com/TimoBolkart/TF_FLAME
Framework tf

AutoLearn - Automated Feature Generation and Selection

Title AutoLearn - Automated Feature Generation and Selection
Authors Ambika Kaul, Saket Maheshwary, Vikram Pudi
Abstract In recent years, the importance of feature engineering has been confirmed by the exceptional performance of deep learning techniques, that automate this task for some applications. For others, feature engineering requires substantial manual effort in designing and selecting features and is often tedious and non-scalable. We present AutoLearn, a regression-based feature learning algorithm. Being data-driven, it requires no domain knowledge and is hence generic. Such a representation is learnt by mining pairwise feature associations, identifying the linear or non-linear relationship between each pair, applying regression and selecting those relationships that are stable and improve the prediction performance. Our experimental evaluation on 18 UC Irvine and 7 Gene expression datasets, across different domains, provides evidence that the features learnt through our model can improve the overall prediction accuracy by 13.28%, compared to original feature space and 5.87% over other top performing models, across 8 different classifiers without using any domain knowledge.
Tasks Automated Feature Engineering, Feature Engineering, Feature Importance
Published 2017-11-17
URL https://ieeexplore.ieee.org/abstract/document/8215494
PDF http://web2py.iiit.ac.in/research_centres/publications/download/inproceedings.pdf.88535e0ea3a74e72.4943444d2d20323031372e706466.pdf
PWC https://paperswithcode.com/paper/autolearn-automated-feature-generation-and
Repo https://github.com/saket-maheshwary/AutoLearn
Framework none

DataStories at SemEval-2017 Task 4: Deep LSTM with Attention for Message-level and Topic-based Sentiment Analysis

Title DataStories at SemEval-2017 Task 4: Deep LSTM with Attention for Message-level and Topic-based Sentiment Analysis
Authors Christos Baziotis, Nikos Pelekis, Christos Doulkeridis
Abstract In this paper we present two deep-learning systems that competed at SemEval-2017 Task 4 {``}Sentiment Analysis in Twitter{''}. We participated in all subtasks for English tweets, involving message-level and topic-based sentiment polarity classification and quantification. We use Long Short-Term Memory (LSTM) networks augmented with two kinds of attention mechanisms, on top of word embeddings pre-trained on a big collection of Twitter messages. Also, we present a text processing tool suitable for social network messages, which performs tokenization, word normalization, segmentation and spell correction. Moreover, our approach uses no hand-crafted features or sentiment lexicons. We ranked 1st (tie) in Subtask A, and achieved very competitive results in the rest of the Subtasks. Both the word embeddings and our text processing tool are available to the research community. |
Tasks Feature Engineering, Sentiment Analysis, Tokenization, Word Embeddings
Published 2017-08-01
URL https://www.aclweb.org/anthology/S17-2126/
PDF https://www.aclweb.org/anthology/S17-2126
PWC https://paperswithcode.com/paper/datastories-at-semeval-2017-task-4-deep-lstm
Repo https://github.com/cbaziotis/ekphrasis
Framework none

Grammatical Error Detection Using Error- and Grammaticality-Specific Word Embeddings

Title Grammatical Error Detection Using Error- and Grammaticality-Specific Word Embeddings
Authors Masahiro Kaneko, Yuya Sakaizawa, Mamoru Komachi
Abstract In this study, we improve grammatical error detection by learning word embeddings that consider grammaticality and error patterns. Most existing algorithms for learning word embeddings usually model only the syntactic context of words so that classifiers treat erroneous and correct words as similar inputs. We address the problem of contextual information by considering learner errors. Specifically, we propose two models: one model that employs grammatical error patterns and another model that considers grammaticality of the target word. We determine grammaticality of n-gram sequence from the annotated error tags and extract grammatical error patterns for word embeddings from large-scale learner corpora. Experimental results show that a bidirectional long-short term memory model initialized by our word embeddings achieved the state-of-the-art accuracy by a large margin in an English grammatical error detection task on the First Certificate in English dataset.
Tasks Grammatical Error Detection, Learning Word Embeddings, Word Embeddings
Published 2017-11-01
URL https://www.aclweb.org/anthology/I17-1005/
PDF https://www.aclweb.org/anthology/I17-1005
PWC https://paperswithcode.com/paper/grammatical-error-detection-using-error-and
Repo https://github.com/kanekomasahiro/grammatical-error-detection
Framework none

How Many Stemmata with Root Degree k?

Title How Many Stemmata with Root Degree k?
Authors Armin Hoenen, Steffen Eger, Ralf Gehrke
Abstract
Tasks
Published 2017-07-01
URL https://www.aclweb.org/anthology/W17-3402/
PDF https://www.aclweb.org/anthology/W17-3402
PWC https://paperswithcode.com/paper/how-many-stemmata-with-root-degree-k
Repo https://github.com/ArminHoenen/KFurcatingRootedGregTrees
Framework none

WILDCAT: Weakly Supervised Learning of Deep ConvNets for Image Classification, Pointwise Localization and Segmentation

Title WILDCAT: Weakly Supervised Learning of Deep ConvNets for Image Classification, Pointwise Localization and Segmentation
Authors Thibaut Durand, Taylor Mordan, Nicolas Thome, Matthieu Cord
Abstract This paper introduces WILDCAT, a deep learning method which jointly aims at aligning image regions for gaining spatial invariance and learning strongly localized features. Our model is trained using only global image labels and is devoted to three main visual recognition tasks: image classification, weakly supervised object localization and semantic segmentation. WILDCAT extends state-of-the-art Convolutional Neural Networks at three main levels: the use of Fully Convolutional Networks for maintaining spatial resolution, the explicit design in the network of local features related to different class modalities, and a new way to pool these features to provide a global image prediction required for weakly supervised training. Extensive experiments show that our model significantly outperforms state-of-the-art methods.
Tasks Image Classification, Object Localization, Semantic Segmentation, Weakly-Supervised Object Localization
Published 2017-07-01
URL http://openaccess.thecvf.com/content_cvpr_2017/html/Durand_WILDCAT_Weakly_Supervised_CVPR_2017_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2017/papers/Durand_WILDCAT_Weakly_Supervised_CVPR_2017_paper.pdf
PWC https://paperswithcode.com/paper/wildcat-weakly-supervised-learning-of-deep
Repo https://github.com/durandtibo/wildcat.pytorch
Framework pytorch

Integrating Semantic Knowledge into Lexical Embeddings Based on Information Content Measurement

Title Integrating Semantic Knowledge into Lexical Embeddings Based on Information Content Measurement
Authors Hsin-Yang Wang, Wei-Yun Ma
Abstract Distributional word representations are widely used in NLP tasks. These representations are based on an assumption that words with a similar context tend to have a similar meaning. To improve the quality of the context-based embeddings, many researches have explored how to make full use of existing lexical resources. In this paper, we argue that while we incorporate the prior knowledge with context-based embeddings, words with different occurrences should be treated differently. Therefore, we propose to rely on the measurement of information content to control the degree of applying prior knowledge into context-based embeddings - different words would have different learning rates when adjusting their embeddings. In the result, we demonstrate that our embeddings get significant improvements on two different tasks: Word Similarity and Analogical Reasoning.
Tasks Document Classification, Information Retrieval, Question Answering, Word Embeddings
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-2082/
PDF https://www.aclweb.org/anthology/E17-2082
PWC https://paperswithcode.com/paper/integrating-semantic-knowledge-into-lexical
Repo https://github.com/hywangntut/KBE
Framework none

A Revisit of Sparse Coding Based Anomaly Detection in Stacked RNN Framework

Title A Revisit of Sparse Coding Based Anomaly Detection in Stacked RNN Framework
Authors Weixin Luo, Wen Liu, Shenghua Gao
Abstract Motivated by the capability of sparse coding based anomaly detection, we propose a Temporally-coherent Sparse Coding (TSC) where we enforce similar neighbouring frames be encoded with similar reconstruction coefficients. Then we map the TSC with a special type of stacked Recurrent Neural Network (sRNN). By taking advantage sRNN in learning all parameters simultaneously, the nontrivial hyper-parameter selection to TSC can be avoided, meanwhile with a shallow sRNN, the reconstruction coefficients can be inferred within a forward pass, which reduces the computational cost for learning sparse coefficients. The contributions of this paper are two-fold: i) We propose a TSC, which can be mapped to a sRNN which facilitates the parameter optimization and accelerates the anomaly prediction. ii) We build a very large dataset which is even larger than the summation of all existing dataset for anomaly detection in terms of both the volume of data and the diversity of scenes. Extensive experiments on both a toy dataset and real datasets demonstrate that our TSC based and sRNN based method consistently outperform existing methods, which validates the effectiveness of our method.
Tasks Anomaly Detection
Published 2017-10-01
URL http://openaccess.thecvf.com/content_iccv_2017/html/Luo_A_Revisit_of_ICCV_2017_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2017/papers/Luo_A_Revisit_of_ICCV_2017_paper.pdf
PWC https://paperswithcode.com/paper/a-revisit-of-sparse-coding-based-anomaly
Repo https://github.com/StevenLiuWen/sRNN_TSC_Anomaly_Detection
Framework tf

Neural system identification for large populations separating “what” and “where”

Title Neural system identification for large populations separating “what” and “where”
Authors David Klindt, Alexander S. Ecker, Thomas Euler, Matthias Bethge
Abstract Neuroscientists classify neurons into different types that perform similar computations at different locations in the visual field. Traditional methods for neural system identification do not capitalize on this separation of “what” and “where”. Learning deep convolutional feature spaces that are shared among many neurons provides an exciting path forward, but the architectural design needs to account for data limitations: While new experimental techniques enable recordings from thousands of neurons, experimental time is limited so that one can sample only a small fraction of each neuron’s response space. Here, we show that a major bottleneck for fitting convolutional neural networks (CNNs) to neural data is the estimation of the individual receptive field locations – a problem that has been scratched only at the surface thus far. We propose a CNN architecture with a sparse readout layer factorizing the spatial (where) and feature (what) dimensions. Our network scales well to thousands of neurons and short recordings and can be trained end-to-end. We evaluate this architecture on ground-truth data to explore the challenges and limitations of CNN-based system identification. Moreover, we show that our network model outperforms current state-of-the art system identification models of mouse primary visual cortex.
Tasks
Published 2017-12-01
URL http://papers.nips.cc/paper/6942-neural-system-identification-for-large-populations-separating-what-and-where
PDF http://papers.nips.cc/paper/6942-neural-system-identification-for-large-populations-separating-what-and-where.pdf
PWC https://paperswithcode.com/paper/neural-system-identification-for-large-1
Repo https://github.com/david-klindt/NIPS2017
Framework tf

Saliency Pattern Detection by Ranking Structured Trees

Title Saliency Pattern Detection by Ranking Structured Trees
Authors Lei Zhu, Haibin Ling, Jin Wu, Huiping Deng, Jin Liu
Abstract In this paper we propose a new salient object detection method via structured label prediction. By learning appearance features in rectangular regions, our structural region representation encodes the local saliency distribution with a matrix of binary labels. We show that the linear combination of structured labels can well model the saliency distribution in local regions. Representing region saliency with structured labels has two advantages: 1) it connects the label assignment of all enclosed pixels, which produces a smooth saliency prediction; and 2) regular-shaped nature of structured labels enables well definition of traditional cues such as regional properties and center surround contrast, and these cues help to build meaningful and informative saliency measures. To measure the consistency between a structured label and the corresponding saliency distribution, we further propose an adaptive label ranking algorithm using proposals that are generated by a CNN model. Finally, we introduce a K-NN enhanced graph representation for saliency propagation, which is more favorable for our task than the widely-used adjacent-graph-based ones. Experimental results demonstrate the effectiveness of our proposed method on six popular benchmarks compared with state-of-the-art approaches.
Tasks Object Detection, Saliency Prediction, Salient Object Detection
Published 2017-10-01
URL http://openaccess.thecvf.com/content_iccv_2017/html/Zhu_Saliency_Pattern_Detection_ICCV_2017_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2017/papers/Zhu_Saliency_Pattern_Detection_ICCV_2017_paper.pdf
PWC https://paperswithcode.com/paper/saliency-pattern-detection-by-ranking
Repo https://github.com/zhulei2016/RST-saliency
Framework none

An open-source tool for negation detection: a maximum-margin approach

Title An open-source tool for negation detection: a maximum-margin approach
Authors Martine Enger, Erik Velldal, Lilja {\O}vrelid
Abstract This paper presents an open-source toolkit for negation detection. It identifies negation cues and their corresponding scope in either raw or parsed text using maximum-margin classification. The system design draws on best practice from the existing literature on negation detection, aiming for a simple and portable system that still achieves competitive performance. Pre-trained models and experimental results are provided for English.
Tasks Negation Detection, Structured Prediction
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-1810/
PDF https://www.aclweb.org/anthology/W17-1810
PWC https://paperswithcode.com/paper/an-open-source-tool-for-negation-detection-a
Repo https://github.com/marenger/negtool
Framework tf

Scattertext: a Browser-Based Tool for Visualizing how Corpora Differ

Title Scattertext: a Browser-Based Tool for Visualizing how Corpora Differ
Authors Jason Kessler
Abstract
Tasks
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-4015/
PDF https://www.aclweb.org/anthology/P17-4015
PWC https://paperswithcode.com/paper/scattertext-a-browser-based-tool-for
Repo https://github.com/JasonKessler/scattertext
Framework tf

Hunt For The Unique, Stable, Sparse And Fast Feature Learning On Graphs

Title Hunt For The Unique, Stable, Sparse And Fast Feature Learning On Graphs
Authors Saurabh Verma, Zhi-Li Zhang
Abstract For the purpose of learning on graphs, we hunt for a graph feature representation that exhibit certain uniqueness, stability and sparsity properties while also being amenable to fast computation. This leads to the discovery of family of graph spectral distances (denoted as FGSD) and their based graph feature representations, which we prove to possess most of these desired properties. To both evaluate the quality of graph features produced by FGSD and demonstrate their utility, we apply them to the graph classification problem. Through extensive experiments, we show that a simple SVM based classification algorithm, driven with our powerful FGSD based graph features, significantly outperforms all the more sophisticated state-of-art algorithms on the unlabeled node datasets in terms of both accuracy and speed; it also yields very competitive results on the labeled datasets - despite the fact it does not utilize any node label information.
Tasks Graph Classification
Published 2017-12-01
URL http://papers.nips.cc/paper/6614-hunt-for-the-unique-stable-sparse-and-fast-feature-learning-on-graphs
PDF http://papers.nips.cc/paper/6614-hunt-for-the-unique-stable-sparse-and-fast-feature-learning-on-graphs.pdf
PWC https://paperswithcode.com/paper/hunt-for-the-unique-stable-sparse-and-fast
Repo https://github.com/benedekrozemberczki/karateclub
Framework none

Font Size: Community Preserving Network Embedding

Title Font Size: Community Preserving Network Embedding
Authors Xiao Wang, Peng Cui, Jing Wang, Jian Pei, Wenwu Zhu, Shiqiang Yang
Abstract Network embedding, aiming to learn the low-dimensional representations of nodes in networks, is of paramount importance in many real applications. One basic requirement of network embedding is to preserve the structure and inherent properties of the networks. While previous network embedding methods primarily preserve the microscopic structure, such as the first- and second-order proximities of nodes, the mesoscopic community structure, which is one of the most prominent feature of networks, is largely ignored. In this paper, we propose a novel Modularized Nonnegative Matrix Factorization (M-NMF) model to incorporate the community structure into network embedding. We exploit the consensus relationship between the representations of nodes and community structure, and then jointly optimize NMF based representation learning model and modularity based community detection model in a unified framework, which enables the learned representations of nodes to preserve both of the microscopic and community structures. We also provide efficient updating rules to infer the parameters of our model, together with the correctness and convergence guarantees. Extensive experimental results on a variety of real-world networks show the superior performance of the proposed method over the state-of-the-arts.
Tasks Community Detection, Network Embedding, Representation Learning
Published 2017-02-10
URL https://aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14589
PDF https://aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14589/13763.pdf
PWC https://paperswithcode.com/paper/font-size-community-preserving-network
Repo https://github.com/benedekrozemberczki/karateclub
Framework none

Automated WordNet Construction Using Word Embeddings

Title Automated WordNet Construction Using Word Embeddings
Authors Mikhail Khodak, Andrej Risteski, Christiane Fellbaum, Sanjeev Arora
Abstract We present a fully unsupervised method for automated construction of WordNets based upon recent advances in distributional representations of sentences and word-senses combined with readily available machine translation tools. The approach requires very few linguistic resources and is thus extensible to multiple target languages. To evaluate our method we construct two 600-word testsets for word-to-synset matching in French and Russian using native speakers and evaluate the performance of our method along with several other recent approaches. Our method exceeds the best language-specific and multi-lingual automated WordNets in F-score for both languages. The databases we construct for French and Russian, both languages without large publicly available manually constructed WordNets, will be publicly released along with the testsets.
Tasks Information Retrieval, Machine Translation, Word Embeddings, Word Sense Induction
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-1902/
PDF https://www.aclweb.org/anthology/W17-1902
PWC https://paperswithcode.com/paper/automated-wordnet-construction-using-word
Repo https://github.com/mkhodak/pawn
Framework none
comments powered by Disqus