October 15, 2019

2470 words 12 mins read

Paper Group NANR 170

Paper Group NANR 170

Explicitly modeling case improves neural dependency parsing. Language Modeling Teaches You More than Translation Does: Lessons Learned Through Auxiliary Syntactic Task Analysis. Listening Comprehension over Argumentative Content. Binarized LSTM Language Model. Dynamic Feature Learning for Partial Face Recognition. INTERPRETATION OF NEURAL NETWORK I …

Explicitly modeling case improves neural dependency parsing

Title Explicitly modeling case improves neural dependency parsing
Authors Clara Vania, Adam Lopez
Abstract Neural dependency parsing models that compose word representations from characters can presumably exploit morphosyntax when making attachment decisions. How much do they know about morphology? We investigate how well they handle morphological case, which is important for parsing. Our experiments on Czech, German and Russian suggest that adding explicit morphological case{—}either oracle or predicted{—}improves neural dependency parsing, indicating that the learned representations in these models do not fully encode the morphological knowledge that they need, and can still benefit from targeted forms of explicit linguistic modeling.
Tasks Dependency Parsing, Multi-Task Learning
Published 2018-11-01
URL https://www.aclweb.org/anthology/W18-5447/
PDF https://www.aclweb.org/anthology/W18-5447
PWC https://paperswithcode.com/paper/explicitly-modeling-case-improves-neural
Repo
Framework

Language Modeling Teaches You More than Translation Does: Lessons Learned Through Auxiliary Syntactic Task Analysis

Title Language Modeling Teaches You More than Translation Does: Lessons Learned Through Auxiliary Syntactic Task Analysis
Authors Kelly Zhang, Samuel Bowman
Abstract Recently, researchers have found that deep LSTMs trained on tasks like machine translation learn substantial syntactic and semantic information about their input sentences, including part-of-speech. These findings begin to shed light on why pretrained representations, like ELMo and CoVe, are so beneficial for neural language understanding models. We still, though, do not yet have a clear understanding of how the choice of pretraining objective affects the type of linguistic information that models learn. With this in mind, we compare four objectives{—}language modeling, translation, skip-thought, and autoencoding{—}on their ability to induce syntactic and part-of-speech information, holding constant the quantity and genre of the training data, as well as the LSTM architecture.
Tasks CCG Supertagging, Language Modelling, Machine Translation, Part-Of-Speech Tagging, Transfer Learning
Published 2018-11-01
URL https://www.aclweb.org/anthology/W18-5448/
PDF https://www.aclweb.org/anthology/W18-5448
PWC https://paperswithcode.com/paper/language-modeling-teaches-you-more-than
Repo
Framework

Listening Comprehension over Argumentative Content

Title Listening Comprehension over Argumentative Content
Authors Shachar Mirkin, Guy Moshkowich, Matan Orbach, Lili Kotlerman, Yoav Kantor, Tamar Lavee, Michal Jacovi, Yonatan Bilu, Ranit Aharonov, Noam Slonim
Abstract This paper presents a task for machine listening comprehension in the argumentation domain and a corresponding dataset in English. We recorded 200 spontaneous speeches arguing for or against 50 controversial topics. For each speech, we formulated a question, aimed at confirming or rejecting the occurrence of potential arguments in the speech. Labels were collected by listening to the speech and marking which arguments were mentioned by the speaker. We applied baseline methods addressing the task, to be used as a benchmark for future work over this dataset. All data used in this work is freely available for research.
Tasks Machine Reading Comprehension, Question Answering, Reading Comprehension, Speech Recognition
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1078/
PDF https://www.aclweb.org/anthology/D18-1078
PWC https://paperswithcode.com/paper/listening-comprehension-over-argumentative
Repo
Framework

Binarized LSTM Language Model

Title Binarized LSTM Language Model
Authors Xuan Liu, Di Cao, Kai Yu
Abstract Long short-term memory (LSTM) language model (LM) has been widely investigated for automatic speech recognition (ASR) and natural language processing (NLP). Although excellent performance is obtained for large vocabulary tasks, tremendous memory consumption prohibits the use of LSTM LM in low-resource devices. The memory consumption mainly comes from the word embedding layer. In this paper, a novel binarized LSTM LM is proposed to address the problem. Words are encoded into binary vectors and other LSTM parameters are further binarized to achieve high memory compression. This is the first effort to investigate binary LSTM for large vocabulary LM. Experiments on both English and Chinese LM and ASR tasks showed that can achieve a compression ratio of 11.3 without any loss of LM and ASR performances and a compression ratio of 31.6 with acceptable minor performance degradation.
Tasks Language Modelling, Speech Recognition
Published 2018-06-01
URL https://www.aclweb.org/anthology/N18-1192/
PDF https://www.aclweb.org/anthology/N18-1192
PWC https://paperswithcode.com/paper/binarized-lstm-language-model
Repo
Framework

Dynamic Feature Learning for Partial Face Recognition

Title Dynamic Feature Learning for Partial Face Recognition
Authors Lingxiao He, Haiqing Li, Qi Zhang, Zhenan Sun
Abstract Partial face recognition (PFR) in unconstrained environment is a very important task, especially in video surveillance, mobile devices, etc. However, a few studies have tackled how to recognize an arbitrary patch of a face image. This study combines Fully Convolutional Network (FCN) with Sparse Representation Classification (SRC) to propose a novel partial face recognition approach, called Dynamic Feature Matching (DFM), to address partial face images regardless of sizes. Based on DFM, we propose a sliding loss to optimize FCN by reducing the intra-variation between a face patch and face images of a subject, which further improves the performance of DFM. The proposed DFM is evaluated on several partial face databases, including LFW, YTF and CASIA-NIR-Distance databases. Experimental results demonstrate the effectiveness and advantages of DFM in comparison with state-of-the-art PFR methods.
Tasks Face Recognition
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/He_Dynamic_Feature_Learning_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/He_Dynamic_Feature_Learning_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/dynamic-feature-learning-for-partial-face
Repo
Framework

INTERPRETATION OF NEURAL NETWORK IS FRAGILE

Title INTERPRETATION OF NEURAL NETWORK IS FRAGILE
Authors Amirata Ghorbani, Abubakar Abid, James Zou
Abstract In order for machine learning to be deployed and trusted in many applications, it is crucial to be able to reliably explain why the machine learning algorithm makes certain predictions. For example, if an algorithm classifies a given pathology image to be a malignant tumor, then the doctor may need to know which parts of the image led the algorithm to this classification. How to interpret black-box predictors is thus an important and active area of research. A fundamental question is: how much can we trust the interpretation itself? In this paper, we show that interpretation of deep learning predictions is extremely fragile in the following sense: two perceptively indistinguishable inputs with the same predicted label can be assigned very different}interpretations. We systematically characterize the fragility of the interpretations generated by several widely-used feature-importance interpretation methods (saliency maps, integrated gradient, and DeepLIFT) on ImageNet and CIFAR-10. Our experiments show that even small random perturbation can change the feature importance and new systematic perturbations can lead to dramatically different interpretations without changing the label. We extend these results to show that interpretations based on exemplars (e.g. influence functions) are similarly fragile. Our analysis of the geometry of the Hessian matrix gives insight on why fragility could be a fundamental challenge to the current interpretation approaches.
Tasks Feature Importance
Published 2018-01-01
URL https://openreview.net/forum?id=H1xJjlbAZ
PDF https://openreview.net/pdf?id=H1xJjlbAZ
PWC https://paperswithcode.com/paper/interpretation-of-neural-network-is-fragile
Repo
Framework

Inferring Shared Attention in Social Scene Videos

Title Inferring Shared Attention in Social Scene Videos
Authors Lifeng Fan, Yixin Chen, Ping Wei, Wenguan Wang, Song-Chun Zhu
Abstract This paper addresses a new problem of inferring shared attention in third-person social scene videos. Shared attention is a phenomenon that two or more individuals simultaneously look at a common target in social scenes. Perceiving and identifying shared attention in videos plays crucial roles in social activities and social scene understanding. We propose a spatial-temporal neural network to detect shared attention intervals in videos and predict shared attention locations in frames. In each video frame, human gaze directions and potential target boxes are two key features for spatially detecting shared attention in the social scene. In temporal domain, a convolutional Long Short- Term Memory network utilizes the temporal continuity and transition constraints to optimize the predicted shared attention heatmap. We collect a new dataset VideoCoAtt from public TV show videos, containing 380 complex video sequences with more than 492,000 frames that include diverse social scenes for shared attention study. Experiments on this dataset show that our model can effectively infer shared attention in videos. We also empirically verify the effectiveness of different components in our model.
Tasks Scene Understanding
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/Fan_Inferring_Shared_Attention_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/Fan_Inferring_Shared_Attention_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/inferring-shared-attention-in-social-scene
Repo
Framework

Proceedings of the First International Workshop on Spatial Language Understanding

Title Proceedings of the First International Workshop on Spatial Language Understanding
Authors
Abstract
Tasks
Published 2018-06-01
URL https://www.aclweb.org/anthology/W18-1400/
PDF https://www.aclweb.org/anthology/W18-1400
PWC https://paperswithcode.com/paper/proceedings-of-the-first-international-2
Repo
Framework

Knowledge Graph Embedding with Hierarchical Relation Structure

Title Knowledge Graph Embedding with Hierarchical Relation Structure
Authors Zhao Zhang, Fuzhen Zhuang, Meng Qu, Fen Lin, Qing He
Abstract The rapid development of knowledge graphs (KGs), such as Freebase and WordNet, has changed the paradigm for AI-related applications. However, even though these KGs are impressively large, most of them are suffering from incompleteness, which leads to performance degradation of AI applications. Most existing researches are focusing on knowledge graph embedding (KGE) models. Nevertheless, those models simply embed entities and relations into latent vectors without leveraging the rich information from the relation structure. Indeed, relations in KGs conform to a three-layer hierarchical relation structure (HRS), i.e., semantically similar relations can make up relation clusters and some relations can be further split into several fine-grained sub-relations. Relation clusters, relations and sub-relations can fit in the top, the middle and the bottom layer of three-layer HRS respectively. To this end, in this paper, we extend existing KGE models TransE, TransH and DistMult, to learn knowledge representations by leveraging the information from the HRS. Particularly, our approach is capable to extend other KGE models. Finally, the experiment results clearly validate the effectiveness of the proposed approach against baselines.
Tasks Graph Embedding, Information Retrieval, Knowledge Base Completion, Knowledge Graph Embedding, Knowledge Graphs, Question Answering
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1358/
PDF https://www.aclweb.org/anthology/D18-1358
PWC https://paperswithcode.com/paper/knowledge-graph-embedding-with-hierarchical
Repo
Framework

WildDash - Creating Hazard-Aware Benchmarks

Title WildDash - Creating Hazard-Aware Benchmarks
Authors Oliver Zendel, Katrin Honauer, Markus Murschitz, Daniel Steininger, Gustavo Fernandez Dominguez
Abstract Test datasets should contain many different challenging aspects so that the robustness and real-world applicability of algorithms can be assessed. In this work, we present a new test dataset for semantic and instance segmentation for the automotive domain. We have conducted a thorough risk analysis to identify situations and aspects that can reduce the output performance for these tasks. Based on this analysis we have designed our new dataset. Meta-information is supplied to mark which individual visual hazards are present in each test case. Furthermore, a new benchmark evaluation method is presented that uses the meta-information to calculate the robustness of a given algorithm with respect to the individual hazards. We show how this new approach allows for a more expressive characterization of algorithm robustness by comparing three baseline algorithms.
Tasks Instance Segmentation, Semantic Segmentation
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/Oliver_Zendel_WildDash_-_Creating_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/Oliver_Zendel_WildDash_-_Creating_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/wilddash-creating-hazard-aware-benchmarks
Repo
Framework

Simplified Corpus with Core Vocabulary

Title Simplified Corpus with Core Vocabulary
Authors Takumi Maruyama, Kazuhide Yamamoto
Abstract
Tasks Text Simplification
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1185/
PDF https://www.aclweb.org/anthology/L18-1185
PWC https://paperswithcode.com/paper/simplified-corpus-with-core-vocabulary
Repo
Framework

Annotating Modality Expressions and Event Factuality for a Japanese Chess Commentary Corpus

Title Annotating Modality Expressions and Event Factuality for a Japanese Chess Commentary Corpus
Authors Suguru Matsuyoshi, Hirotaka Kameko, Yugo Murawaki, Shinsuke Mori
Abstract
Tasks Image Captioning, Text Generation
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1393/
PDF https://www.aclweb.org/anthology/L18-1393
PWC https://paperswithcode.com/paper/annotating-modality-expressions-and-event
Repo
Framework

Fighting Boredom in Recommender Systems with Linear Reinforcement Learning

Title Fighting Boredom in Recommender Systems with Linear Reinforcement Learning
Authors Romain Warlop, Alessandro Lazaric, Jérémie Mary
Abstract A common assumption in recommender systems (RS) is the existence of a best fixed recommendation strategy. Such strategy may be simple and work at the item level (e.g., in multi-armed bandit it is assumed one best fixed arm/item exists) or implement more sophisticated RS (e.g., the objective of A/B testing is to find the best fixed RS and execute it thereafter). We argue that this assumption is rarely verified in practice, as the recommendation process itself may impact the user’s preferences. For instance, a user may get bored by a strategy, while she may gain interest again, if enough time passed since the last time that strategy was used. In this case, a better approach consists in alternating different solutions at the right frequency to fully exploit their potential. In this paper, we first cast the problem as a Markov decision process, where the rewards are a linear function of the recent history of actions, and we show that a policy considering the long-term influence of the recommendations may outperform both fixed-action and contextual greedy policies. We then introduce an extension of the UCRL algorithm ( L IN UCRL ) to effectively balance exploration and exploitation in an unknown environment, and we derive a regret bound that is independent of the number of states. Finally, we empirically validate the model assumptions and the algorithm in a number of realistic scenarios.
Tasks Recommendation Systems
Published 2018-12-01
URL http://papers.nips.cc/paper/7447-fighting-boredom-in-recommender-systems-with-linear-reinforcement-learning
PDF http://papers.nips.cc/paper/7447-fighting-boredom-in-recommender-systems-with-linear-reinforcement-learning.pdf
PWC https://paperswithcode.com/paper/fighting-boredom-in-recommender-systems-with
Repo
Framework

Automatic 3D Indoor Scene Modeling From Single Panorama

Title Automatic 3D Indoor Scene Modeling From Single Panorama
Authors Yang Yang, Shi Jin, Ruiyang Liu, Sing Bing Kang, Jingyi Yu
Abstract We describe a system that automatically extracts 3D geometry of an indoor scene from a single 2D panorama. Our system recovers the spatial layout by finding the floor, walls, and ceiling; it also recovers shapes of typical indoor objects such as furniture. Using sampled perspective sub-views, we extract geometric cues (lines, vanishing points, orientation map, and surface normals) and semantic cues (saliency and object detection information). These cues are used for ground plane estimation and occlusion reasoning. The global spatial layout is inferred through a constraint graph on line segments and planar superpixels. The recovered layout is then used to guide shape estimation of the remaining objects using their normal information. Experiments on synthetic and real datasets show that our approach is state-of-the-art in both accuracy and efficiency. Our system can handle cluttered scenes with complex geometry that are challenging to existing techniques.
Tasks Object Detection
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/Yang_Automatic_3D_Indoor_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/Yang_Automatic_3D_Indoor_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/automatic-3d-indoor-scene-modeling-from
Repo
Framework

Multimodal Named Entity Disambiguation for Noisy Social Media Posts

Title Multimodal Named Entity Disambiguation for Noisy Social Media Posts
Authors Seungwhan Moon, Leonardo Neves, Vitor Carvalho
Abstract We introduce the new Multimodal Named Entity Disambiguation (MNED) task for multimodal social media posts such as Snapchat or Instagram captions, which are composed of short captions with accompanying images. Social media posts bring significant challenges for disambiguation tasks because 1) ambiguity not only comes from polysemous entities, but also from inconsistent or incomplete notations, 2) very limited context is provided with surrounding words, and 3) there are many emerging entities often unseen during training. To this end, we build a new dataset called SnapCaptionsKB, a collection of Snapchat image captions submitted to public and crowd-sourced stories, with named entity mentions fully annotated and linked to entities in an external knowledge base. We then build a deep zeroshot multimodal network for MNED that 1) extracts contexts from both text and image, and 2) predicts correct entity in the knowledge graph embeddings space, allowing for zeroshot disambiguation of entities unseen in training set as well. The proposed model significantly outperforms the state-of-the-art text-only NED models, showing efficacy and potentials of the MNED task.
Tasks Entity Disambiguation, Image Captioning, Knowledge Graph Embeddings, Opinion Mining
Published 2018-07-01
URL https://www.aclweb.org/anthology/P18-1186/
PDF https://www.aclweb.org/anthology/P18-1186
PWC https://paperswithcode.com/paper/multimodal-named-entity-disambiguation-for
Repo
Framework
comments powered by Disqus