October 15, 2019

2216 words 11 mins read

Paper Group NANR 205

Paper Group NANR 205

Constructing a Lexicon of Relational Nouns. Textual Features Indicative of Writing Proficiency in Elementary School Spanish Documents. Link Weight Prediction with Node Embeddings. Tibetan-Chinese Neural Machine Translation based on Syllable Segmentation. Punctuation as Native Language Interference. Twitter Universal Dependency Parsing for African-A …

Constructing a Lexicon of Relational Nouns

Title Constructing a Lexicon of Relational Nouns
Authors Edward Newell, Jackie C.K. Cheung
Abstract
Tasks Natural Language Inference, Question Answering, Relation Extraction
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1537/
PDF https://www.aclweb.org/anthology/L18-1537
PWC https://paperswithcode.com/paper/constructing-a-lexicon-of-relational-nouns
Repo
Framework

Textual Features Indicative of Writing Proficiency in Elementary School Spanish Documents

Title Textual Features Indicative of Writing Proficiency in Elementary School Spanish Documents
Authors Gemma Bel-Enguix, Diana Due{~n}as Ch{'a}vez, Arturo Curiel D{'\i}az
Abstract Childhood acquisition of written language is not straightforward. Writing skills evolve differently depending on external factors, such as the conditions in which children practice their productions and the quality of their instructors{'} guidance. This can be challenging in low-income areas, where schools may struggle to ensure ideal acquisition conditions. Developing computational tools to support the learning process may counterweight negative environmental influences; however, few work exists on the use of information technologies to improve childhood literacy. This work centers around the computational study of Spanish word and syllable structure in documents written by 2nd and 3rd year elementary school students. The studied texts were compared against a corpus of short stories aimed at the same age group, so as to observe whether the children tend to produce similar written patterns as the ones they are expected to interpret at their literacy level. The obtained results show some significant differences between the two kinds of texts, pointing towards possible strategies for the implementation of new education software in support of written language acquisition.
Tasks Language Acquisition
Published 2018-07-01
URL https://www.aclweb.org/anthology/W18-3716/
PDF https://www.aclweb.org/anthology/W18-3716
PWC https://paperswithcode.com/paper/textual-features-indicative-of-writing
Repo
Framework
Title Link Weight Prediction with Node Embeddings
Authors Yuchen Hou, Lawrence B. Holder
Abstract Application of deep learning has been successful in various domains such as im- age recognition, speech recognition and natural language processing. However, the research on its application in graph mining is still in an early stage. Here we present the first generic deep learning approach to the graph link weight prediction problem based on node embeddings. We evaluate this approach with three differ- ent node embedding techniques experimentally and compare its performance with two state-of-the-art non deep learning baseline approaches. Our experiment re- sults suggest that this deep learning approach outperforms the baselines by up to 70% depending on the dataset and embedding technique applied. This approach shows that deep learning can be successfully applied to link weight prediction to improve prediction accuracy.
Tasks Speech Recognition
Published 2018-01-01
URL https://openreview.net/forum?id=ryZ3KCy0W
PDF https://openreview.net/pdf?id=ryZ3KCy0W
PWC https://paperswithcode.com/paper/link-weight-prediction-with-node-embeddings
Repo
Framework

Tibetan-Chinese Neural Machine Translation based on Syllable Segmentation

Title Tibetan-Chinese Neural Machine Translation based on Syllable Segmentation
Authors Wen Lai, Xiaobing Zhao, Wei Bao
Abstract
Tasks Machine Translation
Published 2018-03-01
URL https://www.aclweb.org/anthology/W18-2203/
PDF https://www.aclweb.org/anthology/W18-2203
PWC https://paperswithcode.com/paper/tibetan-chinese-neural-machine-translation
Repo
Framework

Punctuation as Native Language Interference

Title Punctuation as Native Language Interference
Authors Ilia Markov, Vivi Nastase, Carlo Strapparava
Abstract In this paper, we describe experiments designed to explore and evaluate the impact of punctuation marks on the task of native language identification. Punctuation is specific to each language, and is part of the indicators that overtly represent the manner in which each language organizes and conveys information. Our experiments are organized in various set-ups: the usual multi-class classification for individual languages, also considering classification by language groups, across different proficiency levels, topics and even cross-corpus. The results support our hypothesis that punctuation marks are persistent and robust indicators of the native language of the author, which do not diminish in influence even when a high proficiency level in a non-native language is achieved.
Tasks Language Identification, Native Language Identification
Published 2018-08-01
URL https://www.aclweb.org/anthology/C18-1293/
PDF https://www.aclweb.org/anthology/C18-1293
PWC https://paperswithcode.com/paper/punctuation-as-native-language-interference
Repo
Framework

Twitter Universal Dependency Parsing for African-American and Mainstream American English

Title Twitter Universal Dependency Parsing for African-American and Mainstream American English
Authors Su Lin Blodgett, Johnny Wei, Brendan O{'}Connor
Abstract Due to the presence of both Twitter-specific conventions and non-standard and dialectal language, Twitter presents a significant parsing challenge to current dependency parsing tools. We broaden English dependency parsing to handle social media English, particularly social media African-American English (AAE), by developing and annotating a new dataset of 500 tweets, 250 of which are in AAE, within the Universal Dependencies 2.0 framework. We describe our standards for handling Twitter- and AAE-specific features and evaluate a variety of cross-domain strategies for improving parsing with no, or very little, in-domain labeled data, including a new data synthesis approach. We analyze these methods{'} impact on performance disparities between AAE and Mainstream American English tweets, and assess parsing accuracy for specific AAE lexical and syntactic features. Our annotated data and a parsing model are available at: \url{http://slanglab.cs.umass.edu/TwitterAAE/}.
Tasks Dependency Parsing, Information Retrieval, Language Identification, Part-Of-Speech Tagging, Sentiment Analysis
Published 2018-07-01
URL https://www.aclweb.org/anthology/P18-1131/
PDF https://www.aclweb.org/anthology/P18-1131
PWC https://paperswithcode.com/paper/twitter-universal-dependency-parsing-for
Repo
Framework

Optimal Tuning for Divide-and-conquer Kernel Ridge Regression with Massive Data

Title Optimal Tuning for Divide-and-conquer Kernel Ridge Regression with Massive Data
Authors Ganggang Xu, Zuofeng Shang, Guang Cheng
Abstract Divide-and-conquer is a powerful approach for large and massive data analysis. In the nonparameteric regression setting, although various theoretical frameworks have been established to achieve optimality in estimation or hypothesis testing, how to choose the tuning parameter in a practically effective way is still an open problem. In this paper, we propose a data-driven procedure based on divide-and-conquer for selecting the tuning parameters in kernel ridge regression by modifying the popular Generalized Cross-validation (GCV, Wahba, 1990). While the proposed criterion is computationally scalable for massive data sets, it is also shown under mild conditions to be asymptotically optimal in the sense that minimizing the proposed distributed-GCV (dGCV) criterion is equivalent to minimizing the true global conditional empirical loss of the averaged function estimator, extending the existing optimality results of GCV to the divide-and-conquer framework.
Tasks
Published 2018-07-01
URL https://icml.cc/Conferences/2018/Schedule?showEvent=2205
PDF http://proceedings.mlr.press/v80/xu18f/xu18f.pdf
PWC https://paperswithcode.com/paper/optimal-tuning-for-divide-and-conquer-kernel
Repo
Framework

Multiscale Hidden Markov Models For Covariance Prediction

Title Multiscale Hidden Markov Models For Covariance Prediction
Authors João Sedoc, Jordan Rodu, Dean Foster, Lyle Ungar
Abstract This paper presents a novel variant of hierarchical hidden Markov models (HMMs), the multiscale hidden Markov model (MSHMM), and an associated spectral estimation and prediction scheme that is consistent, finds global optima, and is computationally efficient. Our MSHMM is a generative model of multiple HMMs evolving at different rates where the observation is a result of the additive emissions of the HMMs. While estimation is relatively straightforward, prediction for the MSHMM poses a unique challenge, which we address in this paper. Further, we show that spectral estimation of the MSHMM outperforms standard methods of predicting the asset covariance of stock prices, a widely addressed problem that is multiscale, non-stationary, and requires processing huge amounts of data.
Tasks
Published 2018-01-01
URL https://openreview.net/forum?id=ByW5yxgA-
PDF https://openreview.net/pdf?id=ByW5yxgA-
PWC https://paperswithcode.com/paper/multiscale-hidden-markov-models-for
Repo
Framework

NLPZZX at SemEval-2018 Task 1: Using Ensemble Method for Emotion and Sentiment Intensity Determination

Title NLPZZX at SemEval-2018 Task 1: Using Ensemble Method for Emotion and Sentiment Intensity Determination
Authors Zhengxin Zhang, Qimin Zhou, Hao Wu
Abstract In this paper, we put forward a system that competed at SemEval-2018 Task 1: {``}Affect in Tweets{''}. Our system uses a simple yet effective ensemble method which combines several neural network components. We participate in two subtasks for English tweets: EI-reg and V-reg. For two subtasks, different combinations of neural components are examined. For EI-reg, our system achieves an accuracy of 0.727 in Pearson Correlation Coefficient (all instances) and an accuracy of 0.555 in Pearson Correlation Coefficient (0.5-1). For V-reg, the achieved accuracy scores are respectively 0.835 and 0.670 |
Tasks Sentiment Analysis
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-1015/
PDF https://www.aclweb.org/anthology/S18-1015
PWC https://paperswithcode.com/paper/nlpzzx-at-semeval-2018-task-1-using-ensemble
Repo
Framework

Sparse-Complementary Convolution for Efficient Model Utilization on CNNs

Title Sparse-Complementary Convolution for Efficient Model Utilization on CNNs
Authors Chun-Fu (Richard) Chen, Jinwook Oh, Quanfu Fan, Marco Pistoia, Gwo Giun (Chris) Lee
Abstract We introduce an efficient way to increase the accuracy of convolution neural networks (CNNs) based on high model utilization without increasing any computational complexity. The proposed sparse-complementary convolution replaces regular convolution with sparse and complementary shapes of kernels, covering the same receptive field. By the nature of deep learning, high model utilization of a CNN can be achieved with more simpler kernels rather than fewer complex kernels. This simple but insightful model reuses of recent network architectures, ResNet and DenseNet, can provide better accuracy for most classification tasks (CIFAR-10/100 and ImageNet) compared to their baseline models. By simply replacing the convolution of a CNN with our sparse-complementary convolution, at the same FLOPs and parameters, we can improve top-1 accuracy on ImageNet by 0.33% and 0.18% for ResNet-101 and ResNet-152, respectively. A similar accuracy improvement could be gained by increasing the number of layers in those networks by ~1.5x.
Tasks
Published 2018-01-01
URL https://openreview.net/forum?id=HkeJVllRW
PDF https://openreview.net/pdf?id=HkeJVllRW
PWC https://paperswithcode.com/paper/sparse-complementary-convolution-for
Repo
Framework

Geometry Aware Constrained Optimization Techniques for Deep Learning

Title Geometry Aware Constrained Optimization Techniques for Deep Learning
Authors Soumava Kumar Roy, Zakaria Mhammedi, Mehrtash Harandi
Abstract In this paper, we generalize the Stochastic Gradient Descent (SGD) and RMSProp algorithms to the setting of Riemannian optimization. SGD is a popular method for large scale optimization. In particular, it is widely used to train the weights of Deep Neural Networks. However, gradients computed using standard SGD can have large variance, which is detrimental for the convergence rate of the algorithm. Other methods such as RMSProp and ADAM address this issue. Nevertheless, these methods cannot be directly applied to constrained optimization problems. In this paper, we extend some popular optimization algorithm to the Riemannian (constrained) setting. We substantiate our proposed extensions with a range of relevant problems in machine learning such as incremental Principal Component Analysis, computating the Riemannian centroids of SPD matrices, and Deep Metric Learning. We achieve competitive results against the state of the art for fine-grained object recognition datasets.
Tasks Metric Learning, Object Recognition
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/Roy_Geometry_Aware_Constrained_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/Roy_Geometry_Aware_Constrained_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/geometry-aware-constrained-optimization
Repo
Framework

Learning Generative Models with Locally Disentangled Latent Factors

Title Learning Generative Models with Locally Disentangled Latent Factors
Authors Brady Neal, Alex Lamb, Sherjil Ozair, Devon Hjelm, Aaron Courville, Yoshua Bengio, Ioannis Mitliagkas
Abstract One of the most successful techniques in generative models has been decomposing a complicated generation task into a series of simpler generation tasks. For example, generating an image at a low resolution and then learning to refine that into a high resolution image often improves results substantially. Here we explore a novel strategy for decomposing generation for complicated objects in which we first generate latent variables which describe a subset of the observed variables, and then map from these latent variables to the observed space. We show that this allows us to achieve decoupled training of complicated generative models and present both theoretical and experimental results supporting the benefit of such an approach.
Tasks
Published 2018-01-01
URL https://openreview.net/forum?id=rJTGkKxAZ
PDF https://openreview.net/pdf?id=rJTGkKxAZ
PWC https://paperswithcode.com/paper/learning-generative-models-with-locally
Repo
Framework

Dependency Parsing of Code-Switching Data with Cross-Lingual Feature Representations

Title Dependency Parsing of Code-Switching Data with Cross-Lingual Feature Representations
Authors Niko Partanen, Kyungtae Lim, Michael Rie{\ss}ler, Thierry Poibeau
Abstract
Tasks Dependency Parsing
Published 2018-01-01
URL https://www.aclweb.org/anthology/W18-0201/
PDF https://www.aclweb.org/anthology/W18-0201
PWC https://paperswithcode.com/paper/dependency-parsing-of-code-switching-data
Repo
Framework

Adaptive Learning of Local Semantic and Global Structure Representations for Text Classification

Title Adaptive Learning of Local Semantic and Global Structure Representations for Text Classification
Authors Jianyu Zhao, Zhiqiang Zhan, Qichuan Yang, Yang Zhang, Changjian Hu, Zhensheng Li, Liuxin Zhang, Zhiqiang He
Abstract Representation learning is a key issue for most Natural Language Processing (NLP) tasks. Most existing representation models either learn little structure information or just rely on pre-defined structures, leading to degradation of performance and generalization capability. This paper focuses on learning both local semantic and global structure representations for text classification. In detail, we propose a novel Sandwich Neural Network (SNN) to learn semantic and structure representations automatically without relying on parsers. More importantly, semantic and structure information contribute unequally to the text representation at corpus and instance level. To solve the fusion problem, we propose two strategies: Adaptive Learning Sandwich Neural Network (AL-SNN) and Self-Attention Sandwich Neural Network (SA-SNN). The former learns the weights at corpus level, and the latter further combines attention mechanism to assign the weights at instance level. Experimental results demonstrate that our approach achieves competitive performance on several text classification tasks, including sentiment analysis, question type classification and subjectivity classification. Specifically, the accuracies are MR (82.1{%}), SST-5 (50.4{%}), TREC (96{%}) and SUBJ (93.9{%}).
Tasks Representation Learning, Sentiment Analysis, Text Classification
Published 2018-08-01
URL https://www.aclweb.org/anthology/C18-1173/
PDF https://www.aclweb.org/anthology/C18-1173
PWC https://paperswithcode.com/paper/adaptive-learning-of-local-semantic-and
Repo
Framework

DebugSL: An Interactive Tool for Debugging Sentiment Lexicons

Title DebugSL: An Interactive Tool for Debugging Sentiment Lexicons
Authors Andrew Schneider, John Male, Saroja Bhogadhi, Eduard Dragut
Abstract We introduce DebugSL, a visual (Web) debugging tool for sentiment lexicons (SLs). Its core component implements our algorithms for the automatic detection of polarity inconsistencies in SLs. An inconsistency is a set of words and/or word-senses whose polarity assignments cannot all be simultaneously satisfied. DebugSL finds inconsistencies of small sizes in SLs and has a rich user interface which helps users in the correction process. The project source code is available at \url{https://github.com/atschneid/DebugSL} A screencast of DebugSL can be viewed at \url{https://cis.temple.edu/~edragut/DebugSL.webm}
Tasks Sentiment Analysis
Published 2018-06-01
URL https://www.aclweb.org/anthology/N18-5008/
PDF https://www.aclweb.org/anthology/N18-5008
PWC https://paperswithcode.com/paper/debugsl-an-interactive-tool-for-debugging
Repo
Framework
comments powered by Disqus