January 24, 2020

2444 words 12 mins read

Paper Group NANR 190

Paper Group NANR 190

Deep Defocus Map Estimation Using Domain Adaptation. PIE: Pseudo-Invertible Encoder. Whom to Learn From? Graph- vs. Text-based Word Embeddings. VizWiz-Priv: A Dataset for Recognizing the Presence and Purpose of Private Visual Information in Images Taken by Blind People. Graph Classification with Geometric Scattering. A Priori Estimates of the Gener …

Deep Defocus Map Estimation Using Domain Adaptation

Title Deep Defocus Map Estimation Using Domain Adaptation
Authors Junyong Lee, Sungkil Lee, Sunghyun Cho, Seungyong Lee
Abstract In this paper, we propose the first end-to-end convolutional neural network (CNN) architecture, Defocus Map Estimation Network (DMENet), for spatially varying defocus map estimation. To train the network, we produce a novel depth-of-field (DOF) dataset, SYNDOF, where each image is synthetically blurred with a ground-truth depth map. Due to the synthetic nature of SYNDOF, the feature characteristics of images in SYNDOF can differ from those of real defocused photos. To address this gap, we use domain adaptation that transfers the features of real defocused photos into those of synthetically blurred ones. Our DMENet consists of four subnetworks: blur estimation, domain adaptation, content preservation, and sharpness calibration networks. The subnetworks are connected to each other and jointly trained with their corresponding supervisions in an end-to-end manner. Our method is evaluated on publicly available blur detection and blur estimation datasets and the results show the state-of-the-art performance.In this paper, we propose the first end-to-end convolutional neural network (CNN) architecture, Defocus Map Estimation Network (DMENet), for spatially varying defocus map estimation. To train the network, we produce a novel depth-of-field (DOF) dataset, SYNDOF, where each image is synthetically blurred with a ground-truth depth map. Due to the synthetic nature of SYNDOF, the feature characteristics of images in SYNDOF can differ from those of real defocused photos. To address this gap, we use domain adaptation that transfers the features of real defocused photos into those of synthetically blurred ones. Our DMENet consists of four subnetworks: blur estimation, domain adaptation, content preservation, and sharpness calibration networks. The subnetworks are connected to each other and jointly trained with their corresponding supervisions in an end-to-end manner. Our method is evaluated on publicly available blur detection and blur estimation datasets and the results show the state-of-the-art performance.
Tasks Calibration, Domain Adaptation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Lee_Deep_Defocus_Map_Estimation_Using_Domain_Adaptation_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Lee_Deep_Defocus_Map_Estimation_Using_Domain_Adaptation_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/deep-defocus-map-estimation-using-domain
Repo
Framework

PIE: Pseudo-Invertible Encoder

Title PIE: Pseudo-Invertible Encoder
Authors Jan Jetze Beitler, Ivan Sosnovik, Arnold Smeulders
Abstract We consider the problem of information compression from high dimensional data. Where many studies consider the problem of compression by non-invertible trans- formations, we emphasize the importance of invertible compression. We introduce new class of likelihood-based auto encoders with pseudo bijective architecture, which we call Pseudo Invertible Encoders. We provide the theoretical explanation of their principles. We evaluate Gaussian Pseudo Invertible Encoder on MNIST, where our model outperform WAE and VAE in sharpness of the generated images.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=SkgiX2Aqtm
PDF https://openreview.net/pdf?id=SkgiX2Aqtm
PWC https://paperswithcode.com/paper/pie-pseudo-invertible-encoder
Repo
Framework

Whom to Learn From? Graph- vs. Text-based Word Embeddings

Title Whom to Learn From? Graph- vs. Text-based Word Embeddings
Authors Ma{\l}gorzata Salawa, Ant{'o}nio Branco, Ruben Branco, Jo{~a}o Ant{'o}nio Rodrigues, Chakaveh Saedi
Abstract Vectorial representations of meaning can be supported by empirical data from diverse sources and obtained with diverse embedding approaches. This paper aims at screening this experimental space and reports on an assessment of word embeddings supported (i) by data in raw texts vs. in lexical graphs, (ii) by lexical information encoded in association- vs. inference-based graphs, and obtained (iii) by edge reconstruction- vs. matrix factorisation vs. random walk-based graph embedding methods. The results observed with these experiments indicate that the best solutions with graph-based word embeddings are very competitive, consistently outperforming mainstream text-based ones.
Tasks Graph Embedding, Word Embeddings
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-1120/
PDF https://www.aclweb.org/anthology/R19-1120
PWC https://paperswithcode.com/paper/whom-to-learn-from-graph-vs-text-based-word
Repo
Framework

VizWiz-Priv: A Dataset for Recognizing the Presence and Purpose of Private Visual Information in Images Taken by Blind People

Title VizWiz-Priv: A Dataset for Recognizing the Presence and Purpose of Private Visual Information in Images Taken by Blind People
Authors Danna Gurari, Qing Li, Chi Lin, Yinan Zhao, Anhong Guo, Abigale Stangl, Jeffrey P. Bigham
Abstract We introduce the first visual privacy dataset originating from people who are blind in order to better understand their privacy disclosures and to encourage the development of algorithms that can assist in preventing their unintended disclosures. It includes 8,862 regions showing private content across 5,537 images taken by blind people. Of these, 1,403 are paired with questions and 62% of those directly ask about the private content. Experiments demonstrate the utility of this data for predicting whether an image shows private information and whether a question asks about the private content in an image. The dataset is publicly-shared at http://vizwiz.org/data/.
Tasks
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Gurari_VizWiz-Priv_A_Dataset_for_Recognizing_the_Presence_and_Purpose_of_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Gurari_VizWiz-Priv_A_Dataset_for_Recognizing_the_Presence_and_Purpose_of_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/vizwiz-priv-a-dataset-for-recognizing-the
Repo
Framework

Graph Classification with Geometric Scattering

Title Graph Classification with Geometric Scattering
Authors Feng Gao, Guy Wolf, Matthew Hirn
Abstract One of the most notable contributions of deep learning is the application of convolutional neural networks (ConvNets) to structured signal classification, and in particular image classification. Beyond their impressive performances in supervised learning, the structure of such networks inspired the development of deep filter banks referred to as scattering transforms. These transforms apply a cascade of wavelet transforms and complex modulus operators to extract features that are invariant to group operations and stable to deformations. Furthermore, ConvNets inspired recent advances in geometric deep learning, which aim to generalize these networks to graph data by applying notions from graph signal processing to learn deep graph filter cascades. We further advance these lines of research by proposing a geometric scattering transform using graph wavelets defined in terms of random walks on the graph. We demonstrate the utility of features extracted with this designed deep filter bank in graph classification of biochemistry and social network data (incl. state of the art results in the latter case), and in data exploration, where they enable inference of EC exchange preferences in enzyme evolution.
Tasks Graph Classification, Image Classification
Published 2019-05-01
URL https://openreview.net/forum?id=SygK6sA5tX
PDF https://openreview.net/pdf?id=SygK6sA5tX
PWC https://paperswithcode.com/paper/graph-classification-with-geometric
Repo
Framework

A Priori Estimates of the Generalization Error for Two-layer Neural Networks

Title A Priori Estimates of the Generalization Error for Two-layer Neural Networks
Authors Lei Wu, Chao Ma, Weinan E
Abstract New estimates for the generalization error are established for a nonlinear regression problem using a two-layer neural network model. These new estimates are a priori in nature in the sense that the bounds depend only on some norms of the underlying functions to be fitted, not the parameters in the model. In contrast, most existing results for neural networks are a posteriori in nature in the sense that the bounds depend on some norms of the model parameters. The error rates are comparable to that of the Monte Carlo method in terms of the size of the dataset. Moreover, these bounds are equally effective in the over-parametrized regime when the network size is much larger than the size of the dataset.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=Sklqvo0qt7
PDF https://openreview.net/pdf?id=Sklqvo0qt7
PWC https://paperswithcode.com/paper/a-priori-estimates-of-the-generalization
Repo
Framework

ConSSED at SemEval-2019 Task 3: Configurable Semantic and Sentiment Emotion Detector

Title ConSSED at SemEval-2019 Task 3: Configurable Semantic and Sentiment Emotion Detector
Authors Rafa{\l} Po{'s}wiata
Abstract This paper describes our system participating in the SemEval-2019 Task 3: EmoContext: Contextual Emotion Detection in Text. The goal was to for a given textual dialogue, i.e. a user utterance along with two turns of context, identify the emotion of user utterance as one of the emotion classes: Happy, Sad, Angry or Others. Our system: ConSSED is a configurable combination of semantic and sentiment neural models. The official task submission achieved a micro-average F1 score of 75.31 which placed us 16th out of 165 participating systems.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2027/
PDF https://www.aclweb.org/anthology/S19-2027
PWC https://paperswithcode.com/paper/conssed-at-semeval-2019-task-3-configurable
Repo
Framework

M2FPA: A Multi-Yaw Multi-Pitch High-Quality Dataset and Benchmark for Facial Pose Analysis

Title M2FPA: A Multi-Yaw Multi-Pitch High-Quality Dataset and Benchmark for Facial Pose Analysis
Authors Peipei Li, Xiang Wu, Yibo Hu, Ran He, Zhenan Sun
Abstract Facial images in surveillance or mobile scenarios often have large view-point variations in terms of pitch and yaw angles. These jointly occurred angle variations make face recognition challenging. Current public face databases mainly consider the case of yaw variations. In this paper, a new large-scale Multi-yaw Multi-pitch high-quality database is proposed for Facial Pose Analysis (M2FPA), including face frontalization, face rotation, facial pose estimation and pose-invariant face recognition. It contains 397,544 images of 229 subjects with yaw, pitch, attribute, illumination and accessory. M2FPA is the most comprehensive multi-view face database for facial pose analysis. Further, we provide an effective benchmark for face frontalization and pose-invariant face recognition on M2FPA with several state-of-the-art methods, including DR-GAN, TP-GAN and CAPG-GAN. We believe that the new database and benchmark can significantly push forward the advance of facial pose analysis in real-world applications. Moreover, a simple yet effective parsing guided discriminator is introduced to capture the local consistency during GAN optimization. Extensive quantitative and qualitative results on M2FPA and Multi-PIE demonstrate the superiority of our face frontalization method. Baseline results for both face synthesis and face recognition from state-of-the-art methods demonstrate the challenge offered by this new database.
Tasks Face Generation, Face Recognition, Pose Estimation, Robust Face Recognition
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Li_M2FPA_A_Multi-Yaw_Multi-Pitch_High-Quality_Dataset_and_Benchmark_for_Facial_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Li_M2FPA_A_Multi-Yaw_Multi-Pitch_High-Quality_Dataset_and_Benchmark_for_Facial_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/m2fpa-a-multi-yaw-multi-pitch-high-quality-1
Repo
Framework

Electronical resources for Livonian

Title Electronical resources for Livonian
Authors Valts Ern{\v{s}}treits
Abstract
Tasks
Published 2019-01-01
URL https://www.aclweb.org/anthology/W19-0314/
PDF https://www.aclweb.org/anthology/W19-0314
PWC https://paperswithcode.com/paper/electronical-resources-for-livonian
Repo
Framework

Two Generator Game: Learning to Sample via Linear Goodness-of-Fit Test

Title Two Generator Game: Learning to Sample via Linear Goodness-of-Fit Test
Authors Lizhong Ding, Mengyang Yu, Li Liu, Fan Zhu, Yong Liu, Yu Li, Ling Shao
Abstract Learning the probability distribution of high-dimensional data is a challenging problem. To solve this problem, we formulate a deep energy adversarial network (DEAN), which casts the energy model learned from real data into an optimization of a goodness-of-fit (GOF) test statistic. DEAN can be interpreted as a GOF game between two generative networks, where one explicit generative network learns an energy-based distribution that fits the real data, and the other implicit generative network is trained by minimizing a GOF test statistic between the energy-based distribution and the generated data, such that the underlying distribution of the generated data is close to the energy-based distribution. We design a two-level alternative optimization procedure to train the explicit and implicit generative networks, such that the hyper-parameters can also be automatically learned. Experimental results show that DEAN achieves high quality generations compared to the state-of-the-art approaches.
Tasks
Published 2019-12-01
URL http://papers.nips.cc/paper/9304-two-generator-game-learning-to-sample-via-linear-goodness-of-fit-test
PDF http://papers.nips.cc/paper/9304-two-generator-game-learning-to-sample-via-linear-goodness-of-fit-test.pdf
PWC https://paperswithcode.com/paper/two-generator-game-learning-to-sample-via
Repo
Framework

MaxpoolNMS: Getting Rid of NMS Bottlenecks in Two-Stage Object Detectors

Title MaxpoolNMS: Getting Rid of NMS Bottlenecks in Two-Stage Object Detectors
Authors Lile Cai, Bin Zhao, Zhe Wang, Jie Lin, Chuan Sheng Foo, Mohamed Sabry Aly, Vijay Chandrasekhar
Abstract Modern convolutional object detectors have improved the detection accuracy significantly, which in turn inspired the development of dedicated hardware accelerators to achieve real-time performance by exploiting inherent parallelism in the algorithm. Non-maximum suppression (NMS) is an indispensable operation in object detection. In stark contrast to most operations, the commonly-adopted GreedyNMS algorithm does not foster parallelism, which can be a major performance bottleneck. In this paper, we introduce MaxpoolNMS, a parallelizable alternative to the NMS algorithm, which is based on max-pooling classification score maps. By employing a novel multi-scale multi-channel max-pooling strategy, our method is 20x faster than GreedyNMS while simultaneously achieves comparable accuracy, when quantified across various benchmarking datasets, i.e., MS COCO, KITTI and PASCAL VOC. Furthermore, our method is better suited for hardware-based acceleration than GreedyNMS.
Tasks Object Detection
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Cai_MaxpoolNMS_Getting_Rid_of_NMS_Bottlenecks_in_Two-Stage_Object_Detectors_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Cai_MaxpoolNMS_Getting_Rid_of_NMS_Bottlenecks_in_Two-Stage_Object_Detectors_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/maxpoolnms-getting-rid-of-nms-bottlenecks-in
Repo
Framework

Using a Lexical Semantic Network for the Ontology Building

Title Using a Lexical Semantic Network for the Ontology Building
Authors Nadia Bebeshina-Clairet, Sylvie Despres, Mathieu Lafourcade
Abstract Building multilingual ontologies is a hard task as ontologies are often data-rich resources. We introduce an approach which allows exploiting structured lexical semantic knowledge for the ontology building. Given a multilingual lexical semantic (non ontological) resource and an ontology model, it allows mining relevant semantic knowledge and make the ontology building and enhancement process faster.
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-1012/
PDF https://www.aclweb.org/anthology/R19-1012
PWC https://paperswithcode.com/paper/using-a-lexical-semantic-network-for-the
Repo
Framework

BrainEE at SemEval-2019 Task 3: Ensembling Linear Classifiers for Emotion Prediction

Title BrainEE at SemEval-2019 Task 3: Ensembling Linear Classifiers for Emotion Prediction
Authors Vachagan Gratian
Abstract The paper describes an ensemble of linear perceptrons trained for emotion classification as part of the SemEval-2019 shared-task 3. The model uses a matrix of probabilities to weight the activations of the base-classifiers and makes a final prediction using the sum rule. The base-classifiers are multi-class perceptrons utilizing character and word n-grams, part-of-speech tags and sentiment polarity scores. The results of our experiments indicate that the ensemble outperforms the base-classifiers, but only marginally. In the best scenario our model attains an F-Micro score of 0.672, whereas the base-classifiers attained scores ranging from 0.636 to 0.666.
Tasks Emotion Classification
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2020/
PDF https://www.aclweb.org/anthology/S19-2020
PWC https://paperswithcode.com/paper/brainee-at-semeval-2019-task-3-ensembling
Repo
Framework

Generating Fluent Adversarial Examples for Natural Languages

Title Generating Fluent Adversarial Examples for Natural Languages
Authors Huangzhao Zhang, Hao Zhou, Ning Miao, Lei Li
Abstract Efficiently building an adversarial attacker for natural language processing (NLP) tasks is a real challenge. Firstly, as the sentence space is discrete, it is difficult to make small perturbations along the direction of gradients. Secondly, the fluency of the generated examples cannot be guaranteed. In this paper, we propose MHA, which addresses both problems by performing Metropolis-Hastings sampling, whose proposal is designed with the guidance of gradients. Experiments on IMDB and SNLI show that our proposed MHAoutperforms the baseline model on attacking capability. Adversarial training with MHA also leads to better robustness and performance.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1559/
PDF https://www.aclweb.org/anthology/P19-1559
PWC https://paperswithcode.com/paper/generating-fluent-adversarial-examples-for
Repo
Framework

SINAI at SemEval-2019 Task 3: Using affective features for emotion classification in textual conversations

Title SINAI at SemEval-2019 Task 3: Using affective features for emotion classification in textual conversations
Authors Flor Miriam Plaza-del-Arco, M. Dolores Molina-Gonz{'a}lez, Maite Martin, L. Alfonso Ure{~n}a-L{'o}pez
Abstract Detecting emotions in textual conversation is a challenging problem in absence of nonverbal cues typically associated with emotion, like fa- cial expression or voice modulations. How- ever, more and more users are using message platforms such as WhatsApp or Telegram. For this reason, it is important to develop systems capable of understanding human emotions in textual conversations. In this paper, we carried out different systems to analyze the emotions of textual dialogue from SemEval-2019 Task 3: EmoContext for English language. Our main contribution is the integration of emotional and sentimental features in the classification using the SVM algorithm.
Tasks Emotion Classification
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2053/
PDF https://www.aclweb.org/anthology/S19-2053
PWC https://paperswithcode.com/paper/sinai-at-semeval-2019-task-3-using-affective
Repo
Framework
comments powered by Disqus