January 25, 2020

2340 words 11 mins read

Paper Group NANR 36

Paper Group NANR 36

Contextualized Word Representations from Distant Supervision with and for NER. A deep-learning-based approach for fast and robust steel surface defects classification. Self-Supervised 3D Hand Pose Estimation Through Training by Fitting. Crowdsourced Hedge Term Disambiguation. Multilingual Multimodal Machine Translation for Dravidian Languages utili …

Contextualized Word Representations from Distant Supervision with and for NER

Title Contextualized Word Representations from Distant Supervision with and for NER
Authors Abbas Ghaddar, Phillippe Langlais
Abstract We describe a special type of deep contextualized word representation that is learned from distant supervision annotations and dedicated to named entity recognition. Our extensive experiments on 7 datasets show systematic gains across all domains over strong baselines, and demonstrate that our representation is complementary to previously proposed embeddings. We report new state-of-the-art results on CONLL and ONTONOTES datasets.
Tasks Named Entity Recognition
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5513/
PDF https://www.aclweb.org/anthology/D19-5513
PWC https://paperswithcode.com/paper/contextualized-word-representations-from
Repo
Framework

A deep-learning-based approach for fast and robust steel surface defects classification

Title A deep-learning-based approach for fast and robust steel surface defects classification
Authors Guizhong Fu, Peize Sun a, Wenbin Zhu, Jiangxin Yang, Yanlong Cao, Michael Ying Yang, Yanpeng Cao
Abstract Automatic visual recognition of steel surface defects provides critical functionality to facilitate quality control of steel strip production. In this paper, we present a compact yet effective convolutional neural network (CNN) model, which emphasizes the training of low-level features and incorporates multiple receptive fields, to achieve fast and accurate steel surface defect classification. Our proposed method adopts the pre-trained SqueezeNet as the backbone architecture. It only requires a small amount of defect-specific training samples to achieve high- accuracy recognition on a diversity-enhanced testing dataset of steel surface defects which contains severe non- uniform illumination, camera noise, and motion blur. Moreover, our proposed light-weight CNN model can meet the requirement of real-time online inspection, running over 100 fps on a computer equipped with a single NVIDIA TITAN X Graphics Processing Unit (12G memory). Codes and a diversity-enhanced testing dataset will be made publicly available.
Tasks
Published 2019-05-15
URL https://doi.org/10.1016/j.optlaseng.2019.05.005
PDF https://reader.elsevier.com/reader/sd/pii/S0143816619301678?token=55C51B01EE82794EBAB609DE04AB86C09BD2FA1B79CF4A1EC3AA3F3A1D36BAE8C08ABF50F8EC88DD1D709446A2A57A86
PWC https://paperswithcode.com/paper/a-deep-learning-based-approach-for-fast-and
Repo
Framework

Self-Supervised 3D Hand Pose Estimation Through Training by Fitting

Title Self-Supervised 3D Hand Pose Estimation Through Training by Fitting
Authors Chengde Wan, Thomas Probst, Luc Van Gool, Angela Yao
Abstract We present a self-supervision method for 3D hand pose estimation from depth maps. We begin with a neural network initialized with synthesized data and fine-tune it on real but unlabelled depth maps by minimizing a set of data-fitting terms. By approximating the hand surface with a set of spheres, we design a differentiable hand renderer to align estimates by comparing the rendered and input depth maps. In addition, we place a set of priors including a data-driven term to further regulate the estimate’s kinematic feasibility. Our method makes highly accurate estimates comparable to current supervised methods which require large amounts of labelled training samples, thereby advancing state-of-the-art in unsupervised learning for hand pose estimation.
Tasks Hand Pose Estimation, Pose Estimation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Wan_Self-Supervised_3D_Hand_Pose_Estimation_Through_Training_by_Fitting_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Wan_Self-Supervised_3D_Hand_Pose_Estimation_Through_Training_by_Fitting_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/self-supervised-3d-hand-pose-estimation
Repo
Framework

Crowdsourced Hedge Term Disambiguation

Title Crowdsourced Hedge Term Disambiguation
Authors Morgan Ulinski, Julia Hirschberg
Abstract We address the issue of acquiring quality annotations of hedging words and phrases, linguistic phenomenona in which words, sounds, or other constructions are used to express ambiguity or uncertainty. Due to the limited availability of existing corpora annotated for hedging, linguists and other language scientists have been constrained as to the extent they can study this phenomenon. In this paper, we introduce a new method of acquiring hedging annotations via crowdsourcing, based on reformulating the task of labeling hedges as a simple word sense disambiguation task. We also introduce a new hedging corpus we have constructed by applying this method, a collection of forum posts annotated using Amazon Mechanical Turk. We found that the crowdsourced judgments we obtained had an inter-annotator agreement of 92.89{%} (Fleiss{'} Kappa=0.751) and, when comparing a subset of these annotations to an expert-annotated gold standard, an accuracy of 96.65{%}.
Tasks Word Sense Disambiguation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4001/
PDF https://www.aclweb.org/anthology/W19-4001
PWC https://paperswithcode.com/paper/crowdsourced-hedge-term-disambiguation
Repo
Framework

Multilingual Multimodal Machine Translation for Dravidian Languages utilizing Phonetic Transcription

Title Multilingual Multimodal Machine Translation for Dravidian Languages utilizing Phonetic Transcription
Authors Bharathi Raja Chakravarthi, Ruba Priyadharshini, Bernardo Stearns, Arun Jayapal, Sridevy S, Mihael Arcan, Manel Zarrouk, John P McCrae
Abstract
Tasks Machine Translation, Multimodal Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-6809/
PDF https://www.aclweb.org/anthology/W19-6809
PWC https://paperswithcode.com/paper/multilingual-multimodal-machine-translation
Repo
Framework

Mental Fatigue Monitoring using Brain Dynamics Preferences

Title Mental Fatigue Monitoring using Brain Dynamics Preferences
Authors Yuangang Pan, Avinash K Singh, Ivor W. Tsang, Chin-teng Lin
Abstract Driver’s cognitive state of mental fatigue significantly affects driving performance and more importantly public safety. Previous studies leverage the response time (RT) as the metric for mental fatigue and aim at estimating the exact value of RT using electroencephalogram (EEG) signals within a regression model. However, due to the easily corrupted EEG signals and also non-smooth RTs during data collection, regular regression methods generally suffer from poor generalization performance. Considering that human response time is the reflection of brain dynamics preference rather than a single value, a novel model called Brain Dynamic ranking (BDrank) has been proposed. BDrank could learn from brain dynamics preferences using EEG data robustly and preserve the ordering corresponding to RTs. BDrank model is based on the regularized alternative ordinal classification comparing to regular regression based practices. Furthermore, a transition matrix is introduced to characterize the reliability of each channel used in EEG data, which helps in learning brain dynamics preferences only from informative EEG channels. In order to handle large-scale EEG signals~and obtain higher generalization, an online-generalized Expectation Maximum (OnlineGEM) algorithm also has been proposed to update BDrank in an online fashion. Comprehensive empirical analysis on EEG signals from 44 participants shows that BDrank together with OnlineGEM achieves substantial improvements in reliability while simultaneously detecting possible less informative and noisy EEG channels.
Tasks EEG
Published 2019-05-01
URL https://openreview.net/forum?id=B1l9qsA5KQ
PDF https://openreview.net/pdf?id=B1l9qsA5KQ
PWC https://paperswithcode.com/paper/mental-fatigue-monitoring-using-brain
Repo
Framework

UniMelb at SemEval-2019 Task 12: Multi-model combination for toponym resolution

Title UniMelb at SemEval-2019 Task 12: Multi-model combination for toponym resolution
Authors Haonan Li, Minghan Wang, Timothy Baldwin, Martin Tomko, Maria Vasardani
Abstract This paper describes our submission to SemEval-2019 Task 12 on toponym resolution over scientific articles. We train separate NER models for toponym detection over text extracted from tables vs. text from the body of the paper, and train another auxiliary model to eliminate misdetected toponyms. For toponym disambiguation, we use an SVM classifier with hand-engineered features. The best setting achieved a strict micro-F1 score of 80.92{%} and overlap micro-F1 score of 86.88{%} in the toponym detection subtask, ranking 2nd out of 8 teams on F1 score. For toponym disambiguation and end-to-end resolution, we officially ranked 2nd and 3rd, respectively.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2231/
PDF https://www.aclweb.org/anthology/S19-2231
PWC https://paperswithcode.com/paper/unimelb-at-semeval-2019-task-12-multi-model
Repo
Framework

Subversive Toxicity Detection using Sentiment Information

Title Subversive Toxicity Detection using Sentiment Information
Authors Eloi Brassard-Gourdeau, Richard Khoury
Abstract The presence of toxic content has become a major problem for many online communities. Moderators try to limit this problem by implementing more and more refined comment filters, but toxic users are constantly finding new ways to circumvent them. Our hypothesis is that while modifying toxic content and keywords to fool filters can be easy, hiding sentiment is harder. In this paper, we explore various aspects of sentiment detection and their correlation to toxicity, and use our results to implement a toxicity detection tool. We then test how adding the sentiment information helps detect toxicity in three different real-world datasets, and incorporate subversion to these datasets to simulate a user trying to circumvent the system. Our results show sentiment information has a positive impact on toxicity detection.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3501/
PDF https://www.aclweb.org/anthology/W19-3501
PWC https://paperswithcode.com/paper/subversive-toxicity-detection-using-sentiment
Repo
Framework

Proceedings of the First International Workshop on Designing Meaning Representations

Title Proceedings of the First International Workshop on Designing Meaning Representations
Authors
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3300/
PDF https://www.aclweb.org/anthology/W19-3300
PWC https://paperswithcode.com/paper/proceedings-of-the-first-international-5
Repo
Framework

GL at SemEval-2019 Task 5: Identifying hateful tweets with a deep learning approach.

Title GL at SemEval-2019 Task 5: Identifying hateful tweets with a deep learning approach.
Authors Gretel Liz De la Pe{~n}a
Abstract This paper describes the system we developed for SemEval 2019 on Multilingual detection of hate speech against immigrants and women in Twitter (HatEval - Task 5). We use an approach based on an Attention-based Long Short-Term Memory Recurrent Neural Network. In particular, we build a Bidirectional LSTM to extract information from the word embeddings over the sentence, then apply attention over the hidden states to estimate the importance of each word and finally feed this context vector to another LSTM model to get a representation. Then, the output obtained with this model is used to get the prediction of each of the sub-tasks.
Tasks Word Embeddings
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2073/
PDF https://www.aclweb.org/anthology/S19-2073
PWC https://paperswithcode.com/paper/gl-at-semeval-2019-task-5-identifying-hateful
Repo
Framework

LT3 at SemEval-2019 Task 5: Multilingual Detection of Hate Speech Against Immigrants and Women in Twitter (hatEval)

Title LT3 at SemEval-2019 Task 5: Multilingual Detection of Hate Speech Against Immigrants and Women in Twitter (hatEval)
Authors Nina Bauwelinck, Gilles Jacobs, V{'e}ronique Hoste, Els Lefever
Abstract This paper describes our contribution to the SemEval-2019 Task 5 on the detection of hate speech against immigrants and women in Twitter (hatEval). We considered a supervised classification-based approach to detect hate speech in English tweets, which combines a variety of standard lexical and syntactic features with specific features for capturing offensive language. Our experimental results show good classification performance on the training data, but a considerable drop in recall on the held-out test set.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2077/
PDF https://www.aclweb.org/anthology/S19-2077
PWC https://paperswithcode.com/paper/lt3-at-semeval-2019-task-5-multilingual
Repo
Framework

Semantics-Enhanced Adversarial Nets for Text-to-Image Synthesis

Title Semantics-Enhanced Adversarial Nets for Text-to-Image Synthesis
Authors Hongchen Tan, Xiuping Liu, Xin Li, Yi Zhang, Baocai Yin
Abstract This paper presents a new model, Semantics-enhanced Generative Adversarial Network (SEGAN), for fine-grained text-to-image generation. We introduce two modules, a Semantic Consistency Module (SCM) and an Attention Competition Module (ACM), to our SEGAN. The SCM incorporates image-level semantic consistency into the training of the Generative Adversarial Network (GAN), and can diversify the generated images and improve their structural coherence. A Siamese network and two types of semantic similarities are designed to map the synthesized image and the groundtruth image to nearby points in the latent semantic feature space. The ACM constructs adaptive attention weights to differentiate keywords from unimportant words, and improves the stability and accuracy of SEGAN. Extensive experiments demonstrate that our SEGAN significantly outperforms existing state-of-the-art methods in generating photo-realistic images. All source codes and models will be released for comparative study.
Tasks Image Generation, Text-to-Image Generation
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Tan_Semantics-Enhanced_Adversarial_Nets_for_Text-to-Image_Synthesis_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Tan_Semantics-Enhanced_Adversarial_Nets_for_Text-to-Image_Synthesis_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/semantics-enhanced-adversarial-nets-for-text
Repo
Framework

Polynomial Representation for Persistence Diagram

Title Polynomial Representation for Persistence Diagram
Authors Zhichao Wang, Qian Li, Gang Li, Guandong Xu
Abstract Persistence diagram (PD) has been considered as a compact descriptor for topological data analysis (TDA). Unfortunately, PD cannot be directly used in machine learning methods since it is a multiset of points. Recent efforts have been devoted to transforming PDs into vectors to accommodate machine learning methods. However, they share one common shortcoming: the mapping of PDs to a feature representation depends on a pre-defined polynomial. To address this limitation, this paper proposes an algebraic representation for PDs, i.e., polynomial representation. In this work, we discover a set of general polynomials that vanish on vectorized PDs and extract the task-adapted feature representation from these polynomials. We also prove two attractive properties of the proposed polynomial representation, i.e., stability and linear separability. Experiments also show that our method compares favorably with state-of-the-art TDA methods.
Tasks Topological Data Analysis
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Wang_Polynomial_Representation_for_Persistence_Diagram_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Wang_Polynomial_Representation_for_Persistence_Diagram_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/polynomial-representation-for-persistence
Repo
Framework

Scalable Verified Training for Provably Robust Image Classification

Title Scalable Verified Training for Provably Robust Image Classification
Authors Sven Gowal, Krishnamurthy (Dj) Dvijotham, Robert Stanforth, Rudy Bunel, Chongli Qin, Jonathan Uesato, Relja Arandjelovic, Timothy Mann, Pushmeet Kohli
Abstract Recent work has shown that it is possible to train deep neural networks that are provably robust to norm-bounded adversarial perturbations. Most of these methods are based on minimizing an upper bound on the worst-case loss over all possible adversarial perturbations. While these techniques show promise, they often result in difficult optimization procedures that remain hard to scale to larger networks. Through a comprehensive analysis, we show how a simple bounding technique, interval bound propagation (IBP), can be exploited to train large provably robust neural networks that beat the state-of-the-art in verified accuracy. While the upper bound computed by IBP can be quite weak for general networks, we demonstrate that an appropriate loss and clever hyper-parameter schedule allow the network to adapt such that the IBP bound is tight. This results in a fast and stable learning algorithm that outperforms more sophisticated methods and achieves state-of-the-art results on MNIST, CIFAR-10 and SVHN. It also allows us to train the largest model to be verified beyond vacuous bounds on a downscaled version of IMAGENET.
Tasks Image Classification
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Gowal_Scalable_Verified_Training_for_Provably_Robust_Image_Classification_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Gowal_Scalable_Verified_Training_for_Provably_Robust_Image_Classification_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/scalable-verified-training-for-provably
Repo
Framework

HoloPose: Holistic 3D Human Reconstruction In-The-Wild

Title HoloPose: Holistic 3D Human Reconstruction In-The-Wild
Authors Riza Alp Guler, Iasonas Kokkinos
Abstract We introduce HoloPose, a method for holistic monocular 3D human body reconstruction. We first introduce a part-based model for 3D model parameter regression that allows our method to operate in-the-wild, gracefully handling severe occlusions and large pose variation. We further train a multi-task network comprising 2D, 3D and Dense Pose estimation to drive the 3D reconstruction task. For this we introduce an iterative refinement method that aligns the model-based 3D estimates of 2D/3D joint positions and DensePose with their image-based counterparts delivered by CNNs, achieving both model-based, global consistency and high spatial accuracy thanks to the bottom-up CNN processing. We validate our contributions on challenging benchmarks, showing that our method allows us to get both accurate joint and 3D surface estimates while operating at more than 10fps in-the-wild. More information about our approach, including videos and demos is available at http://arielai.com/holopose.
Tasks 3D Reconstruction, Pose Estimation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Guler_HoloPose_Holistic_3D_Human_Reconstruction_In-The-Wild_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Guler_HoloPose_Holistic_3D_Human_Reconstruction_In-The-Wild_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/holopose-holistic-3d-human-reconstruction-in
Repo
Framework
comments powered by Disqus