October 15, 2019

2650 words 13 mins read

Paper Group NANR 46

Paper Group NANR 46

Sentiment Analysis: It’s Complicated!. Statistical shape analysis of simplified neuronal trees. Adversarial Similarity Network for Evaluating Image Alignment in Deep Learning Based Registration. Learning Hierarchical Structures On-The-Fly with a Recurrent-Recursive Model for Sequences. Convex Elicitation of Continuous Properties. Distributed Learni …

Sentiment Analysis: It’s Complicated!

Title Sentiment Analysis: It’s Complicated!
Authors Kian Kenyon-Dean, Eisha Ahmed, Scott Fujimoto, Jeremy Georges-Filteau, Christopher Glasz, Barleen Kaur, Lal, Auguste e, Bh, Shruti eri, Robert Belfer, Nirmal Kanagasabai, Roman Sarrazingendron, Rohit Verma, Derek Ruths
Abstract Sentiment analysis is used as a proxy to measure human emotion, where the objective is to categorize text according to some predefined notion of sentiment. Sentiment analysis datasets are typically constructed with gold-standard sentiment labels, assigned based on the results of manual annotations. When working with such annotations, it is common for dataset constructors to discard {}noisy{''} or {}controversial{''} data where there is significant disagreement on the proper label. In datasets constructed for the purpose of Twitter sentiment analysis (TSA), these controversial examples can compose over 30{%} of the originally annotated data. We argue that the removal of such data is a problematic trend because, when performing real-time sentiment classification of short-text, an automated system cannot know a priori which samples would fall into this category of disputed sentiment. We therefore propose the notion of a {``}complicated{''} class of sentiment to categorize such text, and argue that its inclusion in the short-text sentiment analysis framework will improve the quality of automated sentiment analysis systems as they are implemented in real-world settings. We motivate this argument by building and analyzing a new publicly available TSA dataset of over 7,000 tweets annotated with 5x coverage, named MTSA. Our analysis of classifier performance over our dataset offers insights into sentiment analysis dataset and model design, how current techniques would perform in the real world, and how researchers should handle difficult data. |
Tasks Sentiment Analysis, Twitter Sentiment Analysis
Published 2018-06-01
URL https://www.aclweb.org/anthology/N18-1171/
PDF https://www.aclweb.org/anthology/N18-1171
PWC https://paperswithcode.com/paper/sentiment-analysis-itas-complicated
Repo
Framework

Statistical shape analysis of simplified neuronal trees

Title Statistical shape analysis of simplified neuronal trees
Authors Adam Duncan, Eric Klassen, and Anuj Srivastava
Abstract Neuron morphology plays a central role in characterizing cognitive health and functionality of brain structures. The problem of quantifying neuron shapes and capturing statistical variability of shapes is difficult because neurons differ both in geometry and in topology. This paper develops a mathematical representation of neuronal trees, restricting to the trees that consist of: (1) a main branch viewed as a parameterized curve in R 3 R3 , and (2) some number of secondary branches—also parameterized curves in R 3 R3 —which emanate from the main branch at arbitrary points. It imposes a metric on the representation space, in order to compare neuronal shapes, and to obtain optimal deformations (geodesics) across arbitrary trees. The key idea is to impose certain equivalence relations that allow trees with different geometries and topologies to be compared efficiently. The combinatorial problem of matching side branches across trees is reduced to a linear assignment with well-known efficient solutions. This framework is then applied to comparing, clustering, and classifying neurons using fully automated algorithms. The framework is illustrated on three datasets of neuron reconstructions, specifically showing geodesics paths and cross-validated classification between experimental groups.
Tasks
Published 2018-11-03
URL https://projecteuclid.org/euclid.aoas/1536652959
PDF https://projecteuclid.org/download/pdfview_1/euclid.aoas/1536652959
PWC https://paperswithcode.com/paper/statistical-shape-analysis-of-simplified
Repo
Framework

Adversarial Similarity Network for Evaluating Image Alignment in Deep Learning Based Registration

Title Adversarial Similarity Network for Evaluating Image Alignment in Deep Learning Based Registration
Authors Jingfan Fan,Xiaohuan Cao, Zhong Xue, Pew-Thian Yap, and Dinggang Shen
Abstract This paper introduces an unsupervised adversarial similarity network for image registration. Unlike existing deep learning registration frameworks,our approach does not require ground-truth deformations and specific similarity metrics. We connect a registration network and a discrimination network with adeformable transformation layer. The registration network is trained with feedback from the discrimination network, which is designed to judge whether a pair of registered images are sufficiently similar. Using adversarial training, the registration network is trained to predict deformations that are accurate enough to fool the discrimination network. Experiments on four brain MRI datasets indicate that our method yields registration performance that is promising in both accuracy and efficiency compared with state-of-the-art registration methods, including those based on deep learning.
Tasks Image Registration
Published 2018-09-10
URL https://www.researchgate.net/publication/327628479_Adversarial_Similarity_Network_for_Evaluating_Image_Alignment_in_Deep_Learning_Based_Registration
PDF https://www.researchgate.net/publication/327628479_Adversarial_Similarity_Network_for_Evaluating_Image_Alignment_in_Deep_Learning_Based_Registration
PWC https://paperswithcode.com/paper/adversarial-similarity-network-for-evaluating
Repo
Framework

Learning Hierarchical Structures On-The-Fly with a Recurrent-Recursive Model for Sequences

Title Learning Hierarchical Structures On-The-Fly with a Recurrent-Recursive Model for Sequences
Authors Athul Paul Jacob, Zhouhan Lin, Aless Sordoni, ro, Yoshua Bengio
Abstract We propose a hierarchical model for sequential data that learns a tree on-the-fly, i.e. while reading the sequence. In the model, a recurrent network adapts its structure and reuses recurrent weights in a recursive manner. This creates adaptive skip-connections that ease the learning of long-term dependencies. The tree structure can either be inferred without supervision through reinforcement learning, or learned in a supervised manner. We provide preliminary experiments in a novel Math Expression Evaluation (MEE) task, which is created to have a hierarchical tree structure that can be used to study the effectiveness of our model. Additionally, we test our model in a well-known propositional logic and language modelling tasks. Experimental results have shown the potential of our approach.
Tasks Language Modelling, Representation Learning
Published 2018-07-01
URL https://www.aclweb.org/anthology/W18-3020/
PDF https://www.aclweb.org/anthology/W18-3020
PWC https://paperswithcode.com/paper/learning-hierarchical-structures-on-the-fly
Repo
Framework

Convex Elicitation of Continuous Properties

Title Convex Elicitation of Continuous Properties
Authors Jessica Finocchiaro, Rafael Frongillo
Abstract A property or statistic of a distribution is said to be elicitable if it can be expressed as the minimizer of some loss function in expectation. Recent work shows that continuous real-valued properties are elicitable if and only if they are identifiable, meaning the set of distributions with the same property value can be described by linear constraints. From a practical standpoint, one may ask for which such properties do there exist convex loss functions. In this paper, in a finite-outcome setting, we show that in fact every elicitable real-valued property can be elicited by a convex loss function. Our proof is constructive, and leads to convex loss functions for new properties.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/8241-convex-elicitation-of-continuous-properties
PDF http://papers.nips.cc/paper/8241-convex-elicitation-of-continuous-properties.pdf
PWC https://paperswithcode.com/paper/convex-elicitation-of-continuous-properties
Repo
Framework

Distributed Learning without Distress: Privacy-Preserving Empirical Risk Minimization

Title Distributed Learning without Distress: Privacy-Preserving Empirical Risk Minimization
Authors Bargav Jayaraman, Lingxiao Wang, David Evans, Quanquan Gu
Abstract Distributed learning allows a group of independent data owners to collaboratively learn a model over their data sets without exposing their private data. We present a distributed learning approach that combines differential privacy with secure multi-party computation. We explore two popular methods of differential privacy, output perturbation and gradient perturbation, and advance the state-of-the-art for both methods in the distributed learning setting. In our output perturbation method, the parties combine local models within a secure computation and then add the required differential privacy noise before revealing the model. In our gradient perturbation method, the data owners collaboratively train a global model via an iterative learning algorithm. At each iteration, the parties aggregate their local gradients within a secure computation, adding sufficient noise to ensure privacy before the gradient updates are revealed. For both methods, we show that the noise can be reduced in the multi-party setting by adding the noise inside the secure computation after aggregation, asymptotically improving upon the best previous results. Experiments on real world data sets demonstrate that our methods provide substantial utility gains for typical privacy requirements.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/7871-distributed-learning-without-distress-privacy-preserving-empirical-risk-minimization
PDF http://papers.nips.cc/paper/7871-distributed-learning-without-distress-privacy-preserving-empirical-risk-minimization.pdf
PWC https://paperswithcode.com/paper/distributed-learning-without-distress-privacy
Repo
Framework

Summarizing First-Person Videos from Third Persons’ Points of View

Title Summarizing First-Person Videos from Third Persons’ Points of View
Authors HSUAN-I HO, Wei-Chen Chiu, Yu-Chiang Frank Wang
Abstract Video highlight or summarization is among interesting topics in computer vision, which benefits a variety of applications like viewing, searching, or storage. However, most existing studies rely on training data of third-person videos, which cannot easily generalize to highlight the first-person ones. With the goal of deriving an effective model to summarize first-person videos, we propose a novel deep neural network architecture for describing and discriminating vital spatiotemporal information across videos with different points of view. Our proposed model is realized in a semi-supervised setting, in which fully annotated third-person videos, unlabeled first-person videos, and a small amount of annotated first-person ones are presented during training. In our experiments, qualitative and quantitative evaluations on both benchmarks and our collected first-person video datasets are presented.
Tasks
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/HSUAN-I_HO_Summarizing_First-Person_Videos_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/HSUAN-I_HO_Summarizing_First-Person_Videos_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/summarizing-first-person-videos-from-third-1
Repo
Framework

Proceedings of the 27th International Conference on Computational Linguistics

Title Proceedings of the 27th International Conference on Computational Linguistics
Authors
Abstract
Tasks
Published 2018-08-01
URL https://www.aclweb.org/anthology/C18-1000/
PDF https://www.aclweb.org/anthology/C18-1000
PWC https://paperswithcode.com/paper/proceedings-of-the-27th-international-1
Repo
Framework

Global Convergence of Policy Gradient Methods for Linearized Control Problems

Title Global Convergence of Policy Gradient Methods for Linearized Control Problems
Authors Maryam Fazel, Rong Ge, Sham M. Kakade, Mehran Mesbahi
Abstract Direct policy gradient methods for reinforcement learning and continuous control problems are a popular approach for a variety of reasons: 1) they are easy to implement without explicit knowledge of the underlying model; 2) they are an “end-to-end” approach, directly optimizing the performance metric of interest; 3) they inherently allow for richly parameterized policies. A notable drawback is that even in the most basic continuous control problem (that of linear quadratic regulators), these methods must solve a non-convex optimization problem, where little is understood about their efficiency from both computational and statistical perspectives. In contrast, system identification and model based planning in optimal control theory have a much more solid theoretical footing, where much is known with regards to their computational and statistical properties. This work bridges this gap showing that (model free) policy gradient methods globally converge to the optimal solution and are efficient (polynomially so in relevant problem dependent quantities) with regards to their sample and computational complexities.
Tasks Continuous Control, Policy Gradient Methods
Published 2018-01-01
URL https://openreview.net/forum?id=BJDEbngCZ
PDF https://openreview.net/pdf?id=BJDEbngCZ
PWC https://paperswithcode.com/paper/global-convergence-of-policy-gradient-methods-1
Repo
Framework

Generalization of Learning using Reservoir Computing

Title Generalization of Learning using Reservoir Computing
Authors Sanjukta Krishnagopal, Yiannis Aloimonos, Michelle Girvan
Abstract We investigate the methods by which a Reservoir Computing Network (RCN) learns concepts such as ‘similar’ and ‘different’ between pairs of images using a small training dataset and generalizes these concepts to previously unseen types of data. Specifically, we show that an RCN trained to identify relationships between image-pairs drawn from a subset of digits from the MNIST database or the depth maps of subset of visual scenes from a moving camera generalizes the learned transformations to images of digits unseen during training or depth maps of different visual scenes. We infer, using Principal Component Analysis, that the high dimensional reservoir states generated from an input image pair with a specific transformation converge over time to a unique relationship. Thus, as opposed to training the entire high dimensional reservoir state, the reservoir only needs to train on these unique relationships, allowing the reservoir to perform well with very few training examples. Thus, generalization of learning to unseen images is interpretable in terms of clustering of the reservoir state onto the attractor corresponding to the transformation in reservoir space. We find that RCNs can identify and generalize linear and non-linear transformations, and combinations of transformations, naturally and be a robust and effective image classifier. Additionally, RCNs perform significantly better than state of the art neural network classification techniques such as deep Siamese Neural Networks (SNNs) in generalization tasks both on the MNIST dataset and more complex depth maps of visual scenes from a moving camera. This work helps bridge the gap between explainable machine learning and biological learning through analogies using small datasets, and points to new directions in the investigation of learning processes.
Tasks
Published 2018-01-01
URL https://openreview.net/forum?id=HyFaiGbCW
PDF https://openreview.net/pdf?id=HyFaiGbCW
PWC https://paperswithcode.com/paper/generalization-of-learning-using-reservoir
Repo
Framework

EmoIntens Tracker at SemEval-2018 Task 1: Emotional Intensity Levels in #Tweets

Title EmoIntens Tracker at SemEval-2018 Task 1: Emotional Intensity Levels in #Tweets
Authors Ramona-Andreea Turcu, Amar, S ei, ra Maria, Iuliana-Alex Flescan-Lovin-Arseni, ra, Daniela Gifu, Tr, Diana abat
Abstract The „Affect in Tweets{''} task is centered on emotions categorization and evaluation matrix using multi-language tweets (English and Spanish). In this research, SemEval Affect dataset was preprocessed, categorized, and evaluated accordingly (precision, recall, and accuracy). The system described in this paper is based on the implementation of supervised machine learning (Naive Bayes, KNN and SVM), deep learning (NN Tensor Flow model), and decision trees algorithms.
Tasks Emotion Recognition, Sentiment Analysis
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-1026/
PDF https://www.aclweb.org/anthology/S18-1026
PWC https://paperswithcode.com/paper/emointens-tracker-at-semeval-2018-task-1
Repo
Framework

A Sequence Learning Method for Domain-Specific Entity Linking

Title A Sequence Learning Method for Domain-Specific Entity Linking
Authors Emrah Inan, Oguz Dikenelli
Abstract Recent collective Entity Linking studies usually promote global coherence of all the mapped entities in the same document by using semantic embeddings and graph-based approaches. Although graph-based approaches are shown to achieve remarkable results, they are computationally expensive for general datasets. Also, semantic embeddings only indicate relatedness between entity pairs without considering sequences. In this paper, we address these problems by introducing a two-fold neural model. First, we match easy mention-entity pairs and using the domain information of this pair to filter candidate entities of closer mentions. Second, we resolve more ambiguous pairs using bidirectional Long Short-Term Memory and CRF models for the entity disambiguation. Our proposed system outperforms state-of-the-art systems on the generated domain-specific evaluation dataset.
Tasks Entity Disambiguation, Entity Linking, Knowledge Graph Completion, Link Prediction, Relation Extraction, Word Embeddings
Published 2018-07-01
URL https://www.aclweb.org/anthology/W18-2403/
PDF https://www.aclweb.org/anthology/W18-2403
PWC https://paperswithcode.com/paper/a-sequence-learning-method-for-domain
Repo
Framework

Benchmarks and models for entity-oriented polarity detection

Title Benchmarks and models for entity-oriented polarity detection
Authors Lidia Pivovarova, Arto Klami, Roman Yangarber
Abstract We address the problem of determining entity-oriented polarity in business news. This can be viewed as classifying the polarity of the sentiment expressed toward a given mention of a company in a news article. We present a complete, end-to-end approach to the problem. We introduce a new dataset of over 17,000 manually labeled documents, which is substantially larger than any currently available resources. We propose a benchmark solution based on convolutional neural networks for classifying entity-oriented polarity. Although our dataset is much larger than those currently available, it is small on the scale of datasets commonly used for training robust neural network models. To compensate for this, we use transfer learning{—}pre-train the model on a much larger dataset, annotated for a related but different classification task, in order to learn a good representation for business text, and then fine-tune it on the smaller polarity dataset.
Tasks Sentiment Analysis, Transfer Learning
Published 2018-06-01
URL https://www.aclweb.org/anthology/N18-3016/
PDF https://www.aclweb.org/anthology/N18-3016
PWC https://paperswithcode.com/paper/benchmarks-and-models-for-entity-oriented
Repo
Framework

Yuan at SemEval-2018 Task 1: Tweets Emotion Intensity Prediction using Ensemble Recurrent Neural Network

Title Yuan at SemEval-2018 Task 1: Tweets Emotion Intensity Prediction using Ensemble Recurrent Neural Network
Authors Min Wang, Xiaobing Zhou
Abstract We perform the LSTM and BiLSTM model for the emotion intensity prediction. We only join the third subtask in Task 1:Affect in Tweets. Our system rank 6th among all the teams.
Tasks Sentiment Analysis
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-1031/
PDF https://www.aclweb.org/anthology/S18-1031
PWC https://paperswithcode.com/paper/yuan-at-semeval-2018-task-1-tweets-emotion
Repo
Framework

Large Batch Training of Convolutional Networks with Layer-wise Adaptive Rate Scaling

Title Large Batch Training of Convolutional Networks with Layer-wise Adaptive Rate Scaling
Authors Boris Ginsburg, Igor Gitman, Yang You
Abstract A common way to speed up training of large convolutional networks is to add computational units. Training is then performed using data-parallel synchronous Stochastic Gradient Descent (SGD) with a mini-batch divided between computational units. With an increase in the number of nodes, the batch size grows. However, training with a large batch often results in lower model accuracy. We argue that the current recipe for large batch training (linear learning rate scaling with warm-up) is not general enough and training may diverge. To overcome these optimization difficulties, we propose a new training algorithm based on Layer-wise Adaptive Rate Scaling (LARS). Using LARS, we scaled AlexNet and ResNet-50 to a batch size of 16K.
Tasks
Published 2018-01-01
URL https://openreview.net/forum?id=rJ4uaX2aW
PDF https://openreview.net/pdf?id=rJ4uaX2aW
PWC https://paperswithcode.com/paper/large-batch-training-of-convolutional-1
Repo
Framework
comments powered by Disqus