January 25, 2020

2853 words 14 mins read

Paper Group NANR 17

Paper Group NANR 17

Team Kermit-the-frog at SemEval-2019 Task 4: Bias Detection Through Sentiment Analysis and Simple Linguistic Features. Neural Lemmatization of Multiword Expressions. Collaborative Learning of Semi-Supervised Segmentation and Classification for Medical Images. Integrating Bayesian and Discriminative Sparse Kernel Machines for Multi-class Active Lear …

Team Kermit-the-frog at SemEval-2019 Task 4: Bias Detection Through Sentiment Analysis and Simple Linguistic Features

Title Team Kermit-the-frog at SemEval-2019 Task 4: Bias Detection Through Sentiment Analysis and Simple Linguistic Features
Authors Talita Anthonio, Lennart Kloppenburg
Abstract In this paper we describe our participation in the SemEval 2019 shared task on hyperpartisan news detection. We present the system that we submitted for final evaluation and the three approaches that we used: sentiment, bias-laden words and filtered n-gram features. Our submitted model is a Linear SVM that solely relies on the negative sentiment of a document. We achieved an accuracy of 0.621 and a f1 score of 0.694 in the competition, revealing the predictive power of negative sentiment for this task. There was no major improvement by adding or substituting the features of the other two approaches that we tried.
Tasks Sentiment Analysis
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2177/
PDF https://www.aclweb.org/anthology/S19-2177
PWC https://paperswithcode.com/paper/team-kermit-the-frog-at-semeval-2019-task-4
Repo
Framework

Neural Lemmatization of Multiword Expressions

Title Neural Lemmatization of Multiword Expressions
Authors Marine Schmitt, Mathieu Constant
Abstract This article focuses on the lemmatization of multiword expressions (MWEs). We propose a deep encoder-decoder architecture generating for every MWE word its corresponding part in the lemma, based on the internal context of the MWE. The encoder relies on recurrent networks based on (1) the character sequence of the individual words to capture their morphological properties, and (2) the word sequence of the MWE to capture lexical and syntactic properties. The decoder in charge of generating the corresponding part of the lemma for each word of the MWE is based on a classical character-level attention-based recurrent model. Our model is evaluated for Italian, French, Polish and Portuguese and shows good performances except for Polish.
Tasks Lemmatization
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5117/
PDF https://www.aclweb.org/anthology/W19-5117
PWC https://paperswithcode.com/paper/neural-lemmatization-of-multiword-expressions
Repo
Framework

Collaborative Learning of Semi-Supervised Segmentation and Classification for Medical Images

Title Collaborative Learning of Semi-Supervised Segmentation and Classification for Medical Images
Authors Yi Zhou, Xiaodong He, Lei Huang, Li Liu, Fan Zhu, Shanshan Cui, Ling Shao
Abstract Medical image analysis has two important research areas: disease grading and fine-grained lesion segmentation. Although the former problem often relies on the latter, the two are usually studied separately. Disease severity grading can be treated as a classification problem, which only requires image-level annotations, while the lesion segmentation requires stronger pixel-level annotations. However, pixel-wise data annotation for medical images is highly time-consuming and requires domain experts. In this paper, we propose a collaborative learning method to jointly improve the performance of disease grading and lesion segmentation by semi-supervised learning with an attention mechanism. Given a small set of pixel-level annotated data, a multi-lesion mask generation model first performs the traditional semantic segmentation task. Then, based on initially predicted lesion maps for large quantities of image-level annotated data, a lesion attentive disease grading model is designed to improve the severity classification accuracy. Meanwhile, the lesion attention model can refine the lesion maps using class-specific information to fine-tune the segmentation model in a semi-supervised manner. An adversarial architecture is also integrated for training. With extensive experiments on a representative medical problem called diabetic retinopathy (DR), we validate the effectiveness of our method and achieve consistent improvements over state-of-the-art methods on three public datasets.
Tasks Lesion Segmentation, Semantic Segmentation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Zhou_Collaborative_Learning_of_Semi-Supervised_Segmentation_and_Classification_for_Medical_Images_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Zhou_Collaborative_Learning_of_Semi-Supervised_Segmentation_and_Classification_for_Medical_Images_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/collaborative-learning-of-semi-supervised
Repo
Framework

Integrating Bayesian and Discriminative Sparse Kernel Machines for Multi-class Active Learning

Title Integrating Bayesian and Discriminative Sparse Kernel Machines for Multi-class Active Learning
Authors Weishi Shi, Qi Yu
Abstract We propose a novel active learning (AL) model that integrates Bayesian and discriminative kernel machines for fast and accurate multi-class data sampling. By joining a sparse Bayesian model and a maximum margin machine under a unified kernel machine committee (KMC), the proposed model is able to identify a small number of data samples that best represent the overall data space while accurately capturing the decision boundaries. The integration is conducted using the maximum entropy discrimination framework, resulting in a joint objective function that contains generalized entropy as a regularizer. Such a property allows the proposed AL model to choose data samples that more effectively handle non-separable classification problems. Parameter learning is achieved through a principled optimization framework that leverages convex duality and sparse structure of KMC to efficiently optimize the joint objective function. Key model parameters are used to design a novel sampling function to choose data samples that can simultaneously improve multiple decision boundaries, making it an effective sampler for problems with a large number of classes. Experiments conducted over both synthetic and real data and comparison with competitive AL methods demonstrate the effectiveness of the proposed model.
Tasks Active Learning
Published 2019-12-01
URL http://papers.nips.cc/paper/8500-integrating-bayesian-and-discriminative-sparse-kernel-machines-for-multi-class-active-learning
PDF http://papers.nips.cc/paper/8500-integrating-bayesian-and-discriminative-sparse-kernel-machines-for-multi-class-active-learning.pdf
PWC https://paperswithcode.com/paper/integrating-bayesian-and-discriminative
Repo
Framework

Proceedings of the Shared Task on Cross-Framework Meaning Representation Parsing at the 2019 Conference on Natural Language Learning

Title Proceedings of the Shared Task on Cross-Framework Meaning Representation Parsing at the 2019 Conference on Natural Language Learning
Authors
Abstract
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/K19-2000/
PDF https://www.aclweb.org/anthology/K19-2000
PWC https://paperswithcode.com/paper/proceedings-of-the-shared-task-on-cross
Repo
Framework

EEG Signal Dimensionality Reduction and Classification using Tensor Decomposition and Deep Convolutional Neural Networks

Title EEG Signal Dimensionality Reduction and Classification using Tensor Decomposition and Deep Convolutional Neural Networks
Authors Mojtaba Taherisadr, Mohsen Joneidi, Nazanin Rahnavard
Abstract A new deep learning-based electroencephalography (EEG) signal analysis framework is proposed. While deep neural networks, specifically convolutional neural networks (CNNs), have gained remarkable attention recently, they still suffer from high dimensionality of the training data. Two-dimensional input images of CNNs are more vulnerable to be redundant versus one-dimensional input time-series of conventional neural networks. In this study, we propose a new dimensionality reduction framework for reducing the dimension of CNN inputs based on the tensor decomposition of the time-frequency representation of EEG signals. The proposed tensor decomposition-based dimensionality reduction algorithm transforms a large set of slices of the input tensor to a concise set of slices which are called super-slices. Employing super-slices not only handles the artifacts and redundancies of the EEG data but also reduces the dimension of the CNNs training inputs. We also consider different time-frequency representation methods for EEG image generation and provide a comprehensive comparison among them. We test our proposed framework on HCB-MIT data and as results show our approach outperforms other previous studies.
Tasks Dimensionality Reduction, EEG, Image Generation, Seizure Detection, Time Series
Published 2019-08-27
URL https://arxiv.org/abs/1908.10432
PDF https://arxiv.org/pdf/1908.10432
PWC https://paperswithcode.com/paper/eeg-signal-dimensionality-reduction-and
Repo
Framework

Sequential Attention with Keyword Mask Model for Community-based Question Answering

Title Sequential Attention with Keyword Mask Model for Community-based Question Answering
Authors Jianxin Yang, Wenge Rong, Libin Shi, Zhang Xiong
Abstract In Community-based Question Answering system(CQA), Answer Selection(AS) is a critical task, which focuses on finding a suitable answer within a list of candidate answers. For neural network models, the key issue is how to model the representations of QA text pairs and calculate the interactions between them. We propose a Sequential Attention with Keyword Mask model(SAKM) for CQA to imitate human reading behavior. Question and answer text regard each other as context within keyword-mask attention when encoding the representations, and repeat multiple times(hops) in a sequential style. So the QA pairs capture features and information from both question text and answer text, interacting and improving vector representations iteratively through hops. The flexibility of the model allows to extract meaningful keywords from the sentences and enhance diverse mutual information. We perform on answer selection tasks and multi-level answer ranking tasks. Experiment results demonstrate the superiority of our proposed model on community-based QA datasets.
Tasks Answer Selection, Question Answering
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-1228/
PDF https://www.aclweb.org/anthology/N19-1228
PWC https://paperswithcode.com/paper/sequential-attention-with-keyword-mask-model
Repo
Framework

On Tighter Generalization Bounds for Deep Neural Networks: CNNs, ResNets, and Beyond

Title On Tighter Generalization Bounds for Deep Neural Networks: CNNs, ResNets, and Beyond
Authors Xingguo Li, Junwei Lu, Zhaoran Wang, Jarvis Haupt, Tuo Zhao
Abstract We propose a generalization error bound for a general family of deep neural networks based on the depth and width of the networks, as well as the spectral norm of weight matrices. Through introducing a novel characterization of the Lipschitz properties of neural network family, we achieve a tighter generalization error bound. We further obtain a result that is free of linear dependence on norms for bounded losses. Besides the general deep neural networks, our results can be applied to derive new bounds for several popular architectures, including convolutional neural networks (CNNs), residual networks (ResNets), and hyperspherical networks (SphereNets). When achieving same generalization errors with previous arts, our bounds allow for the choice of much larger parameter spaces of weight matrices, inducing potentially stronger expressive ability for neural networks.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=SJzwvoCqF7
PDF https://openreview.net/pdf?id=SJzwvoCqF7
PWC https://paperswithcode.com/paper/on-tighter-generalization-bounds-for-deep
Repo
Framework

Automatic Taxonomy Induction and Expansion

Title Automatic Taxonomy Induction and Expansion
Authors Nicolas Rodolfo Fauceglia, Alfio Gliozzo, Sarthak Dash, Md. Faisal Mahbub Chowdhury, N Mihindukulasooriya, ana
Abstract The Knowledge Graph Induction Service (KGIS) is an end-to-end knowledge induction system. One of its main capabilities is to automatically induce taxonomies from input documents using a hybrid approach that takes advantage of linguistic patterns, semantic web and neural networks. KGIS allows the user to semi-automatically curate and expand the induced taxonomy through a component called Smart SpreadSheet by exploiting distributional semantics. In this paper, we describe these taxonomy induction and expansion features of KGIS. A screencast video demonstrating the system is available in https://ibm.box.com/v/emnlp-2019-demo .
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-3005/
PDF https://www.aclweb.org/anthology/D19-3005
PWC https://paperswithcode.com/paper/automatic-taxonomy-induction-and-expansion
Repo
Framework

Learning from Bad Data via Generation

Title Learning from Bad Data via Generation
Authors Tianyu Guo, Chang Xu, Boxin Shi, Chao Xu, Dacheng Tao
Abstract Bad training data would challenge the learning model from understanding the underlying data-generating scheme, which then increases the difficulty in achieving satisfactory performance on unseen test data. We suppose the real data distribution lies in a distribution set supported by the empirical distribution of bad data. A worst-case formulation can be developed over this distribution set, and then be interpreted as a generation task in an adversarial manner. The connections and differences between GANs and our framework have been thoroughly discussed. We further theoretically show the influence of this generation task on learning from bad data and reveal its connection with a data-dependent regularization. Given different distance measures (\eg, Wasserstein distance or JS divergence) of distributions, we can derive different objective functions for the problem. Experimental results on different kinds of bad training data demonstrate the necessity and effectiveness of the proposed method.
Tasks
Published 2019-12-01
URL http://papers.nips.cc/paper/8837-learning-from-bad-data-via-generation
PDF http://papers.nips.cc/paper/8837-learning-from-bad-data-via-generation.pdf
PWC https://paperswithcode.com/paper/learning-from-bad-data-via-generation
Repo
Framework

Low Level Linguistic Controls for Style Transfer and Content Preservation

Title Low Level Linguistic Controls for Style Transfer and Content Preservation
Authors Katy Gero, Chris Kedzie, Jonathan Reeve, Lydia Chilton
Abstract Despite the success of style transfer in image processing, it has seen limited progress in natural language generation. Part of the problem is that content is not as easily decoupled from style in the text domain. Curiously, in the field of stylometry, content does not figure prominently in practical methods of discriminating stylistic elements, such as authorship and genre. Rather, syntax and function words are the most salient features. Drawing on this work, we model style as a suite of low-level linguistic controls, such as frequency of pronouns, prepositions, and subordinate clause constructions. We train a neural encoder-decoder model to reconstruct reference sentences given only content words and the setting of the controls. We perform style transfer by keeping the content words fixed while adjusting the controls to be indicative of another style. In experiments, we show that the model reliably responds to the linguistic controls and perform both automatic and manual evaluations on style transfer. We find we can fool a style classifier 84{%} of the time, and that our model produces highly diverse and stylistically distinctive outputs. This work introduces a formal, extendable model of style that can add control to any neural text generation system.
Tasks Style Transfer, Text Generation
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-8628/
PDF https://www.aclweb.org/anthology/W19-8628
PWC https://paperswithcode.com/paper/low-level-linguistic-controls-for-style-1
Repo
Framework

Spatial Residual Layer and Dense Connection Block Enhanced Spatial Temporal Graph Convolutional Network for Skeleton-Based Action Recognition

Title Spatial Residual Layer and Dense Connection Block Enhanced Spatial Temporal Graph Convolutional Network for Skeleton-Based Action Recognition
Authors Cong Wu, Xiao-Jun Wu, Josef Kittler
Abstract Recent research has shown that modeling the dynamic joint features of the human body by a graph convolutional network (GCN) is a groundbreaking approach for skeleton-based action recognition, especially for the recognition of the body-motion, human-object and human-human interactions. Nevertheless, how to model and utilize coherent skeleton information comprehensively is still an open problem. In order to capture the rich spatiotemporal information and utilize features more effectively, we introduce a spatial residual layer and a dense connection block enhanced spatial temporal graph convolutional network. More specifically, our work introduces three aspects. Firstly, we extend spatial graph convolution to spatial temporal graph convolution of cross-domain residual to extract more precise and informative spatiotemporal feature, and reduce the training complexity by feature fusion in the, so-called, spatial residual layer. Secondly, instead of simply superimposing multiple similar layers, we use dense connection to take full advantage of the global information. Thirdly, we combine the above mentioned two components to create a spatial temporal graph convolutional network (ST-GCN), referred to as SDGCN. The proposed graph representation has a new structure. We perform extensive experiments on two large datasets: Kinetics and NTU-RGB+D. Our method achieves a great improvement in performance compared to the mainstream methods. We evaluate our method quantitatively and qualitatively, thus proving its effectiveness.
Tasks Skeleton Based Action Recognition
Published 2019-10-27
URL http://openaccess.thecvf.com/content_ICCVW_2019/html/SGRL/Wu_Spatial_Residual_Layer_and_Dense_Connection_Block_Enhanced_Spatial_Temporal_ICCVW_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCVW_2019/papers/SGRL/Wu_Spatial_Residual_Layer_and_Dense_Connection_Block_Enhanced_Spatial_Temporal_ICCVW_2019_paper.pdf
PWC https://paperswithcode.com/paper/spatial-residual-layer-and-dense-connection
Repo
Framework

MMAct: A Large-Scale Dataset for Cross Modal Human Action Understanding

Title MMAct: A Large-Scale Dataset for Cross Modal Human Action Understanding
Authors Quan Kong, Ziming Wu, Ziwei Deng, Martin Klinkigt, Bin Tong, Tomokazu Murakami
Abstract Unlike vision modalities, body-worn sensors or passive sensing can avoid the failure of action understanding in vision related challenges, e.g. occlusion and appearance variation. However, a standard large-scale dataset does not exist, in which different types of modalities across vision and sensors are integrated. To address the disadvantage of vision-based modalities and push towards multi/cross modal action understanding, this paper introduces a new large-scale dataset recorded from 20 distinct subjects with seven different types of modalities: RGB videos, keypoints, acceleration, gyroscope, orientation, Wi-Fi and pressure signal. The dataset consists of more than 36k video clips for 37 action classes covering a wide range of daily life activities such as desktop-related and check-in-based ones in four different distinct scenarios. On the basis of our dataset, we propose a novel multi modality distillation model with attention mechanism to realize an adaptive knowledge transfer from sensor-based modalities to vision-based modalities. The proposed model significantly improves performance of action recognition compared to models trained with only RGB information. The experimental results confirm the effectiveness of our model on cross-subject, -view, -scene and -session evaluation criteria. We believe that this new large-scale multimodal dataset will contribute the community of multimodal based action understanding.
Tasks Transfer Learning
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Kong_MMAct_A_Large-Scale_Dataset_for_Cross_Modal_Human_Action_Understanding_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Kong_MMAct_A_Large-Scale_Dataset_for_Cross_Modal_Human_Action_Understanding_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/mmact-a-large-scale-dataset-for-cross-modal
Repo
Framework

Three-Stream Convolutional Neural Network With Multi-Task and Ensemble Learning for 3D Action Recognition

Title Three-Stream Convolutional Neural Network With Multi-Task and Ensemble Learning for 3D Action Recognition
Authors Duohan Liang, Guoliang Fan, Guangfeng Lin, Wanjun Chen, Xiaorong Pan, Hong Zhu
Abstract In this paper, we propose a three-stream convolutional neural network (3SCNN) for action recognition from skeleton sequences, which aims to thoroughly and fully exploit the skeleton data by extracting, learning, fusing and inferring multiple motion-related features, including 3D joint positions and joint displacements across adjacent frames as well as oriented bone segments. The proposed 3SCNN involves three sequential stages. The first stage enriches three independently extracted features by co-occurrence feature learning. The second stage involves multi-channel pairwise fusion to take advantage of the complementary and diverse nature among three features. The third stage is a multi-task and ensemble learning network to further improve the generalization ability of 3SCNN. Experimental results on the standard dataset show the effectiveness of our proposed multi-stream feature learning, fusion and inference method for skeleton-based 3D action recognition.
Tasks 3D Human Action Recognition, Skeleton Based Action Recognition
Published 2019-06-16
URL http://openaccess.thecvf.com/content_CVPRW_2019/html/PBVS/Liang_Three-Stream_Convolutional_Neural_Network_With_Multi-Task_and_Ensemble_Learning_for_CVPRW_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPRW_2019/papers/PBVS/Liang_Three-Stream_Convolutional_Neural_Network_With_Multi-Task_and_Ensemble_Learning_for_CVPRW_2019_paper.pdf
PWC https://paperswithcode.com/paper/three-stream-convolutional-neural-network
Repo
Framework

Machine Translation from an Intercomprehension Perspective

Title Machine Translation from an Intercomprehension Perspective
Authors Yu Chen, Tania Avgustinova
Abstract Within the first shared task on machine translation between similar languages, we present our first attempts on Czech to Polish machine translation from an intercomprehension perspective. We propose methods based on the mutual intelligibility of the two languages, taking advantage of their orthographic and phonological similarity, in the hope to improve over our baselines. The translation results are evaluated using BLEU. On this metric, none of our proposals could outperform the baselines on the final test set. The current setups are rather preliminary, and there are several potential improvements we can try in the future.
Tasks Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5425/
PDF https://www.aclweb.org/anthology/W19-5425
PWC https://paperswithcode.com/paper/machine-translation-from-an
Repo
Framework
comments powered by Disqus