January 25, 2020

2509 words 12 mins read

Paper Group NANR 7

Paper Group NANR 7

SyntaxFest 2019 Invited talk - Arguments and adjuncts. Reading Turn by Turn: Hierarchical Attention Architecture for Spoken Dialogue Comprehension. ConlluEditor: a fully graphical editor for Universal dependencies treebank files. Global Feature Guided Local Pooling. Multiclass Performance Metric Elicitation. Do Nuclear Submarines Have Nuclear Capta …

SyntaxFest 2019 Invited talk - Arguments and adjuncts

Title SyntaxFest 2019 Invited talk - Arguments and adjuncts
Authors Adam Przepi{'o}rkowski
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-8001/
PDF https://www.aclweb.org/anthology/W19-8001
PWC https://paperswithcode.com/paper/syntaxfest-2019-invited-talk-arguments-and
Repo
Framework

Reading Turn by Turn: Hierarchical Attention Architecture for Spoken Dialogue Comprehension

Title Reading Turn by Turn: Hierarchical Attention Architecture for Spoken Dialogue Comprehension
Authors Zhengyuan Liu, Nancy Chen
Abstract Comprehending multi-turn spoken conversations is an emerging research area, presenting challenges different from reading comprehension of passages due to the interactive nature of information exchange from at least two speakers. Unlike passages, where sentences are often the default semantic modeling unit, in multi-turn conversations, a turn is a topically coherent unit embodied with immediately relevant context, making it a linguistically intuitive segment for computationally modeling verbal interactions. Therefore, in this work, we propose a hierarchical attention neural network architecture, combining turn-level and word-level attention mechanisms, to improve spoken dialogue comprehension performance. Experiments are conducted on a multi-turn conversation dataset, where nurses inquire and discuss symptom information with patients. We empirically show that the proposed approach outperforms standard attention baselines, achieves more efficient learning outcomes, and is more robust to lengthy and out-of-distribution test samples.
Tasks Reading Comprehension
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1543/
PDF https://www.aclweb.org/anthology/P19-1543
PWC https://paperswithcode.com/paper/reading-turn-by-turn-hierarchical-attention
Repo
Framework

ConlluEditor: a fully graphical editor for Universal dependencies treebank files

Title ConlluEditor: a fully graphical editor for Universal dependencies treebank files
Authors Johannes Heinecke
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-8010/
PDF https://www.aclweb.org/anthology/W19-8010
PWC https://paperswithcode.com/paper/conllueditor-a-fully-graphical-editor-for
Repo
Framework

Global Feature Guided Local Pooling

Title Global Feature Guided Local Pooling
Authors Takumi Kobayashi
Abstract In deep convolutional neural networks (CNNs), local pooling operation is a key building block to effectively downsize feature maps for reducing computation cost as well as increasing robustness against input variation. There are several types of pooling operation, such as average/max-pooling, from which one has to be manually selected for building CNNs. The optimal pooling type would be dependent on characteristics of features in CNNs and classification tasks, making it hard to find out the proper pooling module in advance. In this paper, we propose a flexible pooling method which adaptively tunes the pooling functionality based on input features without manually fixing it beforehand. In the proposed method, the parameterized pooling form is derived from a probabilistic perspective to flexibly represent various types of pooling and then the parameters are estimated by means of global statistics in the input feature map. Thus, the proposed local pooling guided by global features effectively works in the CNNs trained in an end-to-end manner. The experimental results on image classification tasks demonstrate the effectiveness of the proposed pooling method in various deep CNNs.
Tasks Image Classification
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Kobayashi_Global_Feature_Guided_Local_Pooling_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Kobayashi_Global_Feature_Guided_Local_Pooling_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/global-feature-guided-local-pooling
Repo
Framework

Multiclass Performance Metric Elicitation

Title Multiclass Performance Metric Elicitation
Authors Gaurush Hiranandani, Shant Boodaghians, Ruta Mehta, Oluwasanmi O. Koyejo
Abstract Metric Elicitation is a principled framework for selecting the performance metric that best reflects implicit user preferences. However, available strategies have so far been limited to binary classification. In this paper, we propose novel strategies for eliciting multiclass classification performance metrics using only relative preference feedback. We also show that the strategies are robust to both finite sample and feedback noise.
Tasks
Published 2019-12-01
URL http://papers.nips.cc/paper/9133-multiclass-performance-metric-elicitation
PDF http://papers.nips.cc/paper/9133-multiclass-performance-metric-elicitation.pdf
PWC https://paperswithcode.com/paper/multiclass-performance-metric-elicitation
Repo
Framework

Do Nuclear Submarines Have Nuclear Captains? A Challenge Dataset for Commonsense Reasoning over Adjectives and Objects

Title Do Nuclear Submarines Have Nuclear Captains? A Challenge Dataset for Commonsense Reasoning over Adjectives and Objects
Authors James Mullenbach, Jonathan Gordon, Nanyun Peng, Jonathan May
Abstract How do adjectives project from a noun to its parts? If a motorcycle is red, are its wheels red? Is a nuclear submarine{'}s captain nuclear? These questions are easy for humans to judge using our commonsense understanding of the world, but are difficult for computers. To attack this challenge, we crowdsource a set of human judgments that answer the English-language question {``}Given a whole described by an adjective, does the adjective also describe a given part?{''} We build strong baselines for this task with a classification approach. Our findings indicate that, despite the recent successes of large language models on tasks aimed to assess commonsense knowledge, these models do not greatly outperform simple word-level models based on pre-trained word embeddings. This provides evidence that the amount of commonsense knowledge encoded in these language models does not extend far beyond that already baked into the word embeddings. Our dataset will serve as a useful testbed for future research in commonsense reasoning, especially as it relates to adjectives and objects |
Tasks Word Embeddings
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1625/
PDF https://www.aclweb.org/anthology/D19-1625
PWC https://paperswithcode.com/paper/do-nuclear-submarines-have-nuclear-captains-a
Repo
Framework

Evaluation of Morphological Embeddings for English and Russian Languages

Title Evaluation of Morphological Embeddings for English and Russian Languages
Authors Vitaly Romanov, Albina Khusainova
Abstract This paper evaluates morphology-based embeddings for English and Russian languages. Despite the interest and introduction of several morphology based word embedding models in the past and acclaimed performance improvements on word similarity and language modeling tasks, in our experiments, we did not observe any stable preference over two of our baseline models - SkipGram and FastText. The performance exhibited by morphological embeddings is the average of the two baselines mentioned above.
Tasks Language Modelling
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-2010/
PDF https://www.aclweb.org/anthology/W19-2010
PWC https://paperswithcode.com/paper/evaluation-of-morphological-embeddings-for
Repo
Framework

Align, Attend and Locate: Chest X-Ray Diagnosis via Contrast Induced Attention Network With Limited Supervision

Title Align, Attend and Locate: Chest X-Ray Diagnosis via Contrast Induced Attention Network With Limited Supervision
Authors Jingyu Liu, Gangming Zhao, Yu Fei, Ming Zhang, Yizhou Wang, Yizhou Yu
Abstract Obstacles facing accurate identification and localization of diseases in chest X-ray images lie in the lack of high-quality images and annotations. In this paper, we propose a Contrast Induced Attention Network (CIA-Net), which exploits the highly structured property of chest X-ray images and localizes diseases via contrastive learning on the aligned positive and negative samples. To force the attention module to focus only on sites of abnormalities, we also introduce a learnable alignment module to adjust all the input images, which eliminates variations of scales, angles, and displacements of X-ray images generated under bad scan conditions. We show that the use of contrastive attention and alignment module allows the model to learn rich identification and localization information using only a small amount of location annotations, resulting in state-of-the-art performance in NIH chest X-ray dataset.
Tasks
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Liu_Align_Attend_and_Locate_Chest_X-Ray_Diagnosis_via_Contrast_Induced_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Liu_Align_Attend_and_Locate_Chest_X-Ray_Diagnosis_via_Contrast_Induced_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/align-attend-and-locate-chest-x-ray-diagnosis
Repo
Framework

Complex Program Induction for Querying Knowledge Bases in the Absence of Gold Programs

Title Complex Program Induction for Querying Knowledge Bases in the Absence of Gold Programs
Authors Amrita Saha, Ghulam Ahmed Ansari, Abhishek Laddha, Karthik Sankaranarayanan, Soumen Chakrabarti
Abstract Recent years have seen increasingly complex question-answering on knowledge bases (KBQA) involving logical, quantitative, and comparative reasoning over KB subgraphs. Neural Program Induction (NPI) is a pragmatic approach toward modularizing the reasoning process by translating a complex natural language query into a multi-step executable program. While NPI has been commonly trained with the {}{}gold{'}{'} program or its sketch, for realistic KBQA applications such gold programs are expensive to obtain. There, practically only natural language queries and the corresponding answers can be provided for training. The resulting combinatorial explosion in program space, along with extremely sparse rewards, makes NPI for KBQA ambitious and challenging. We present Complex Imperative Program Induction from Terminal Rewards (CIPITR), an advanced neural programmer that mitigates reward sparsity with auxiliary rewards, and restricts the program space to semantically correct programs using high-level constraints, KB schema, and inferred answer type. CIPITR solves complex KBQA considerably more accurately than key-value memory networks and neural symbolic machines (NSM). For moderately complex queries requiring 2- to 5-step programs, CIPITR scores at least 3{\mbox{$\times$}} higher F1 than the competing systems. On one of the hardest class of programs (comparative reasoning) with 5{–}10 steps, CIPITR outperforms NSM by a factor of 89 and memory networks by 9 times.
Tasks Question Answering
Published 2019-03-01
URL https://www.aclweb.org/anthology/Q19-1012/
PDF https://www.aclweb.org/anthology/Q19-1012
PWC https://paperswithcode.com/paper/complex-program-induction-for-querying
Repo
Framework

DeepSentiPeer: Harnessing Sentiment in Review Texts to Recommend Peer Review Decisions

Title DeepSentiPeer: Harnessing Sentiment in Review Texts to Recommend Peer Review Decisions
Authors Tirthankar Ghosal, Rajeev Verma, Asif Ekbal, Pushpak Bhattacharyya
Abstract Automatically validating a research artefact is one of the frontiers in Artificial Intelligence (AI) that directly brings it close to competing with human intellect and intuition. Although criticised sometimes, the existing peer review system still stands as the benchmark of research validation. The present-day peer review process is not straightforward and demands profound domain knowledge, expertise, and intelligence of human reviewer(s), which is somewhat elusive with the current state of AI. However, the peer review texts, which contains rich sentiment information of the reviewer, reflecting his/her overall attitude towards the research in the paper, could be a valuable entity to predict the acceptance or rejection of the manuscript under consideration. Here in this work, we investigate the role of reviewer sentiment embedded within peer review texts to predict the peer review outcome. Our proposed deep neural architecture takes into account three channels of information: the paper, the corresponding reviews, and review{'}s polarity to predict the overall recommendation score as well as the final decision. We achieve significant performance improvement over the baselines (∼ 29{%} error reduction) proposed in a recently released dataset of peer reviews. An AI of this kind could assist the editors/program chairs as an additional layer of confidence, especially when non-responding/missing reviewers are frequent in present day peer review.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1106/
PDF https://www.aclweb.org/anthology/P19-1106
PWC https://paperswithcode.com/paper/deepsentipeer-harnessing-sentiment-in-review
Repo
Framework

SME-Net: Sparse Motion Estimation for Parametric Video Prediction Through Reinforcement Learning

Title SME-Net: Sparse Motion Estimation for Parametric Video Prediction Through Reinforcement Learning
Authors Yung-Han Ho, Chuan-Yuan Cho, Wen-Hsiao Peng, Guo-Lun Jin
Abstract This paper leverages a classic prediction technique, known as parametric overlapped block motion compensation (POBMC), in a reinforcement learning framework for video prediction. Learning-based prediction methods with explicit motion models often suffer from having to estimate large numbers of motion parameters with artificial regularization. Inspired by the success of sparse motion-based prediction for video compression, we propose a parametric video prediction on a sparse motion field composed of few critical pixels and their motion vectors. The prediction is achieved by gradually refining the estimate of a future frame in iterative, discrete steps. Along the way, the identification of critical pixels and their motion estimation are addressed by two neural networks trained under a reinforcement learning setting. Our model achieves the state-of-the-art performance on CaltchPed, UCF101 and CIF datasets in one-step and multi-step prediction tests. It shows good generalization results and is able to learn well on small training data.
Tasks Motion Compensation, Motion Estimation, Video Compression, Video Prediction
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Ho_SME-Net_Sparse_Motion_Estimation_for_Parametric_Video_Prediction_Through_Reinforcement_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Ho_SME-Net_Sparse_Motion_Estimation_for_Parametric_Video_Prediction_Through_Reinforcement_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/sme-net-sparse-motion-estimation-for
Repo
Framework

Proceedings of the Third Workshop on Structured Prediction for NLP

Title Proceedings of the Third Workshop on Structured Prediction for NLP
Authors
Abstract
Tasks Structured Prediction
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-1500/
PDF https://www.aclweb.org/anthology/W19-1500
PWC https://paperswithcode.com/paper/proceedings-of-the-third-workshop-on-5
Repo
Framework

A Novel Deep Arrhythmia-Diagnosis Network for Atrial Fibrillation Classification Using Electrocardiogram Signals

Title A Novel Deep Arrhythmia-Diagnosis Network for Atrial Fibrillation Classification Using Electrocardiogram Signals
Authors Hao Dang, Muyi Sun, Guanhong Zhang, Xingqun Qi, Xiaoguang Zhou, Qing Chang
Abstract Atrial fibrillation (AF), a common abnormal heartbeat rhythm, is a life-threatening recurrent disease that affects older adults. Automatic classification is one of the most valuable topics in medical sciences and bioinformatics, especially the detection of atrial fibrillation. However, it is difficult to accurately explain the local characteristics of electrocardiogram (ECG) signals by manual analysis, due to their small amplitude and short duration, coupled with the complexity and non-linearity. Hence, in this paper, we propose a novel deep arrhythmia-diagnosis method, named deep CNN-BLSTM network model, to automatically detect the AF heartbeats using the ECG signals. The model mainly consists of four convolution layers: two BLSTM layers and two fully connected layers. The datasets of RR intervals (called set A) and heartbeat sequences (P-QRS-T waves, called set B) are fed into the above-mentioned model. Most importantly, our proposed approach achieved favorable performances with an accuracy of 99.94% and 98.63% in the training and validation set of set A, respectively. In the testing set (unseen data sets), we obtained an accuracy of 96.59%, a sensitivity of 99.93%, and a specificity of 97.03%. To the best of our knowledge, the algorithm we proposed has shown excellent results compared to many state-of-art researches, which provides a new solution for the AF automatic detection.
Tasks Arrhythmia Detection, Atrial Fibrillation Detection, Electrocardiography (ECG)
Published 2019-05-24
URL https://doi.org/10.1109/ACCESS.2019.2918792
PDF https://ieeexplore.ieee.org/ielx7/6287639/8600701/08721643.pdf
PWC https://paperswithcode.com/paper/a-novel-deep-arrhythmia-diagnosis-network-for
Repo
Framework

Variational Sparse Coding

Title Variational Sparse Coding
Authors Francesco Tonolini, Bjorn Sand Jensen, Roderick Murray-Smith
Abstract Variational auto-encoders (VAEs) offer a tractable approach when performing approximate inference in otherwise intractable generative models. However, standard VAEs often produce latent codes that are disperse and lack interpretability, thus making the resulting representations unsuitable for auxiliary tasks (e.g. classification) and human interpretation. We address these issues by merging ideas from variational auto-encoders and sparse coding, and propose to explicitly model sparsity in the latent space of a VAE with a Spike and Slab prior distribution. We derive the evidence lower bound using a discrete mixture recognition function thereby making approximate posterior inference as computational efficient as in the standard VAE case. With the new approach, we are able to infer truly sparse representations with generally intractable non-linear probabilistic models. We show that these sparse representations are advantageous over standard VAE representations on two benchmark classification tasks (MNIST and Fashion-MNIST) by demonstrating improved classification accuracy and significantly increased robustness to the number of latent dimensions. Furthermore, we demonstrate qualitatively that the sparse elements capture subjectively understandable sources of variation.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=SkeJ6iR9Km
PDF https://openreview.net/pdf?id=SkeJ6iR9Km
PWC https://paperswithcode.com/paper/variational-sparse-coding
Repo
Framework

On the Margin Theory of Feedforward Neural Networks

Title On the Margin Theory of Feedforward Neural Networks
Authors Colin Wei, Jason Lee, Qiang Liu, Tengyu Ma
Abstract Past works have shown that, somewhat surprisingly, over-parametrization can help generalization in neural networks. Towards explaining this phenomenon, we adopt a margin-based perspective. We establish: 1) for multi-layer feedforward relu networks, the global minimizer of a weakly-regularized cross-entropy loss has the maximum normalized margin among all networks, 2) as a result, increasing the over-parametrization improves the normalized margin and generalization error bounds for deep networks. In the case of two-layer networks, an infinite-width neural network enjoys the best generalization guarantees. The typical infinite feature methods are kernel methods; we compare the neural net margin with that of kernel methods and construct natural instances where kernel methods have much weaker generalization guarantees. We validate this gap between the two approaches empirically. Finally, this infinite-neuron viewpoint is also fruitful for analyzing optimization. We show that a perturbed gradient flow on infinite-size networks finds a global optimizer in polynomial time.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=HJGtFoC5Fm
PDF https://openreview.net/pdf?id=HJGtFoC5Fm
PWC https://paperswithcode.com/paper/on-the-margin-theory-of-feedforward-neural-1
Repo
Framework
comments powered by Disqus