October 16, 2019

2271 words 11 mins read

Paper Group NAWR 20

Paper Group NAWR 20

Encoder-Decoder Methods for Text Normalization. Greedy Hash: Towards Fast Optimization for Accurate Hash Coding in CNN. Data-driven computing in elasticity via kernel regression. Cross-lingual Knowledge Graph Alignment via Graph Convolutional Networks. Multi-value Rule Sets for Interpretable Classification with Feature-Efficient Representations. St …

Encoder-Decoder Methods for Text Normalization

Title Encoder-Decoder Methods for Text Normalization
Authors Massimo Lusetti, Tatyana Ruzsics, Anne G{"o}hring, Tanja Samard{\v{z}}i{'c}, Elisabeth Stark
Abstract Text normalization is the task of mapping non-canonical language, typical of speech transcription and computer-mediated communication, to a standardized writing. It is an up-stream task necessary to enable the subsequent direct employment of standard natural language processing tools and indispensable for languages such as Swiss German, with strong regional variation and no written standard. Text normalization has been addressed with a variety of methods, most successfully with character-level statistical machine translation (CSMT). In the meantime, machine translation has changed and the new methods, known as neural encoder-decoder (ED) models, resulted in remarkable improvements. Text normalization, however, has not yet followed. A number of neural methods have been tried, but CSMT remains the state-of-the-art. In this work, we normalize Swiss German WhatsApp messages using the ED framework. We exploit the flexibility of this framework, which allows us to learn from the same training data in different ways. In particular, we modify the decoding stage of a plain ED model to include target-side language models operating at different levels of granularity: characters and words. Our systematic comparison shows that our approach results in an improvement over the CSMT state-of-the-art.
Tasks Machine Translation
Published 2018-08-01
URL https://www.aclweb.org/anthology/W18-3902/
PDF https://www.aclweb.org/anthology/W18-3902
PWC https://paperswithcode.com/paper/encoder-decoder-methods-for-text
Repo https://github.com/tatyana-ruzsics/uzh-corpuslab-normalization
Framework none

Greedy Hash: Towards Fast Optimization for Accurate Hash Coding in CNN

Title Greedy Hash: Towards Fast Optimization for Accurate Hash Coding in CNN
Authors Shupeng Su, Chao Zhang, Kai Han, Yonghong Tian
Abstract To convert the input into binary code, hashing algorithm has been widely used for approximate nearest neighbor search on large-scale image sets due to its computation and storage efficiency. Deep hashing further improves the retrieval quality by combining the hash coding with deep neural network. However, a major difficulty in deep hashing lies in the discrete constraints imposed on the network output, which generally makes the optimization NP hard. In this work, we adopt the greedy principle to tackle this NP hard problem by iteratively updating the network toward the probable optimal discrete solution in each iteration. A hash coding layer is designed to implement our approach which strictly uses the sign function in forward propagation to maintain the discrete constraints, while in back propagation the gradients are transmitted intactly to the front layer to avoid the vanishing gradients. In addition to the theoretical derivation, we provide a new perspective to visualize and understand the effectiveness and efficiency of our algorithm. Experiments on benchmark datasets show that our scheme outperforms state-of-the-art hashing methods in both supervised and unsupervised tasks.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/7360-greedy-hash-towards-fast-optimization-for-accurate-hash-coding-in-cnn
PDF http://papers.nips.cc/paper/7360-greedy-hash-towards-fast-optimization-for-accurate-hash-coding-in-cnn.pdf
PWC https://paperswithcode.com/paper/greedy-hash-towards-fast-optimization-for
Repo https://github.com/ssppp/GreedyHash
Framework pytorch

Data-driven computing in elasticity via kernel regression

Title Data-driven computing in elasticity via kernel regression
Authors Yoshihiro Kanno
Abstract This paper presents a simple nonparametric regression approach to data-driven computing in elasticity. We apply the kernel regression to the material data set, and formulate a system of nonlinear equations solved to obtain a static equilibrium state of an elastic structure. Preliminary numerical experiments illustrate that, compared with existing methods, the proposed method finds a reasonable solution even if data points distribute coarsely in a given material data set.
Tasks Stress-Strain Relation
Published 2018-12-12
URL https://www.sciencedirect.com/science/article/pii/S2095034918302071
PDF https://www.sciencedirect.com/science/article/pii/S2095034918302071
PWC https://paperswithcode.com/paper/data-driven-computing-in-elasticity-via-1
Repo https://github.com/ykanno22/data_driven_kernel_regression
Framework none

Cross-lingual Knowledge Graph Alignment via Graph Convolutional Networks

Title Cross-lingual Knowledge Graph Alignment via Graph Convolutional Networks
Authors Zhichun Wang, Qingsong Lv, Xiaohan Lan, Yu Zhang
Abstract Multilingual knowledge graphs (KGs) such as DBpedia and YAGO contain structured knowledge of entities in several distinct languages, and they are useful resources for cross-lingual AI and NLP applications. Cross-lingual KG alignment is the task of matching entities with their counterparts in different languages, which is an important way to enrich the cross-lingual links in multilingual KGs. In this paper, we propose a novel approach for cross-lingual KG alignment via graph convolutional networks (GCNs). Given a set of pre-aligned entities, our approach trains GCNs to embed entities of each language into a unified vector space. Entity alignments are discovered based on the distances between entities in the embedding space. Embeddings can be learned from both the structural and attribute information of entities, and the results of structure embedding and attribute embedding are combined to get accurate alignments. In the experiments on aligning real multilingual KGs, our approach gets the best performance compared with other embedding-based KG alignment approaches.
Tasks Feature Engineering, Knowledge Graphs, Machine Translation
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1032/
PDF https://www.aclweb.org/anthology/D18-1032
PWC https://paperswithcode.com/paper/cross-lingual-knowledge-graph-alignment-via
Repo https://github.com/1049451037/GCN-Align
Framework tf

Multi-value Rule Sets for Interpretable Classification with Feature-Efficient Representations

Title Multi-value Rule Sets for Interpretable Classification with Feature-Efficient Representations
Authors Tong Wang
Abstract We present the Multi-value Rule Set (MRS) for interpretable classification with feature efficient presentations. Compared to rule sets built from single-value rules, MRS adopts a more generalized form of association rules that allows multiple values in a condition. Rules of this form are more concise than classical single-value rules in capturing and describing patterns in data. Our formulation also pursues a higher efficiency of feature utilization, which reduces possible cost in data collection and storage. We propose a Bayesian framework for formulating an MRS model and develop an efficient inference method for learning a maximum a posteriori, incorporating theoretically grounded bounds to iteratively reduce the search space and improve the search efficiency. Experiments on synthetic and real-world data demonstrate that MRS models have significantly smaller complexity and fewer features than baseline models while being competitive in predictive accuracy.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/8281-multi-value-rule-sets-for-interpretable-classification-with-feature-efficient-representations
PDF http://papers.nips.cc/paper/8281-multi-value-rule-sets-for-interpretable-classification-with-feature-efficient-representations.pdf
PWC https://paperswithcode.com/paper/multi-value-rule-sets-for-interpretable
Repo https://github.com/wangtongada/MRS
Framework none

Stimulus domain transfer in recurrent models for large scale cortical population prediction on video

Title Stimulus domain transfer in recurrent models for large scale cortical population prediction on video
Authors Fabian Sinz, Alexander S. Ecker, Paul Fahey, Edgar Walker, Erick Cobos, Emmanouil Froudarakis, Dimitri Yatsenko, Zachary Pitkow, Jacob Reimer, Andreas Tolias
Abstract To better understand the representations in visual cortex, we need to generate better predictions of neural activity in awake animals presented with their ecological input: natural video. Despite recent advances in models for static images, models for predicting responses to natural video are scarce and standard linear-nonlinear models perform poorly. We developed a new deep recurrent network architecture that predicts inferred spiking activity of thousands of mouse V1 neurons simultaneously recorded with two-photon microscopy, while accounting for confounding factors such as the animal’s gaze position and brain state changes related to running state and pupil dilation. Powerful system identification models provide an opportunity to gain insight into cortical functions through in silico experiments that can subsequently be tested in the brain. However, in many cases this approach requires that the model is able to generalize to stimulus statistics that it was not trained on, such as band-limited noise and other parameterized stimuli. We investigated these domain transfer properties in our model and find that our model trained on natural images is able to correctly predict the orientation tuning of neurons in responses to artificial noise stimuli. Finally, we show that we can fully generalize from movies to noise and maintain high predictive performance on both stimulus domains by fine-tuning only the final layer’s weights on a network otherwise trained on natural movies. The converse, however, is not true.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/7950-stimulus-domain-transfer-in-recurrent-models-for-large-scale-cortical-population-prediction-on-video
PDF http://papers.nips.cc/paper/7950-stimulus-domain-transfer-in-recurrent-models-for-large-scale-cortical-population-prediction-on-video.pdf
PWC https://paperswithcode.com/paper/stimulus-domain-transfer-in-recurrent-models
Repo https://github.com/sinzlab/Sinz2018_NIPS
Framework pytorch

Urdu Word Embeddings

Title Urdu Word Embeddings
Authors Samar Haider
Abstract
Tasks Semantic Textual Similarity, Word Embeddings
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1155/
PDF https://www.aclweb.org/anthology/L18-1155
PWC https://paperswithcode.com/paper/urdu-word-embeddings
Repo https://github.com/samarh/urduvec
Framework none

ETPC - A Paraphrase Identification Corpus Annotated with Extended Paraphrase Typology and Negation

Title ETPC - A Paraphrase Identification Corpus Annotated with Extended Paraphrase Typology and Negation
Authors Venelin Kovatchev, M. Ant{`o}nia Mart{'\i}, Maria Salam{'o}
Abstract
Tasks Natural Language Inference, Paraphrase Identification, Question Answering, Semantic Textual Similarity, Text Simplification, Text Summarization
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1221/
PDF https://www.aclweb.org/anthology/L18-1221
PWC https://paperswithcode.com/paper/etpc-a-paraphrase-identification-corpus
Repo https://github.com/venelink/ETPC
Framework none

Semi-Orthogonal Low-Rank Matrix Factorization for Deep Neural Networks

Title Semi-Orthogonal Low-Rank Matrix Factorization for Deep Neural Networks
Authors Daniel Povey, Gaofeng Cheng, Yiming Wang, Ke Li, Hainan Xu, Mahsa Yarmohammadi, Sanjeev Khudanpur
Abstract Time Delay Neural Networks (TDNNs), also known as onedimensional Convolutional Neural Networks (1-d CNNs), are an efficient and well-performing neural network architecture for speech recognition. We introduce a factored form of TDNNs (TDNN-F) which is structurally the same as a TDNN whose layers have been compressed via SVD, but is trained from a random start with one of the two factors of each matrix constrained to be semi-orthogonal. This gives substantial improvements over TDNNs and performs about as well as TDNN-LSTM hybrids.
Tasks Speech Recognition
Published 2018-09-02
URL https://www.isca-speech.org/archive/Interspeech_2018/abstracts/1417.html
PDF https://www.danielpovey.com/files/2018_interspeech_tdnnf.pdf
PWC https://paperswithcode.com/paper/semi-orthogonal-low-rank-matrix-factorization
Repo https://github.com/cvqluu/Factorized-TDNN
Framework pytorch

Bidirectional Feature Pyramid Network with Recurrent Attention Residual Modules for Shadow Detection

Title Bidirectional Feature Pyramid Network with Recurrent Attention Residual Modules for Shadow Detection
Authors Lei Zhu, Zijun Deng, Xiaowei Hu, Chi-Wing Fu, Xuemiao Xu, Jing Qin, Pheng-Ann Heng
Abstract This paper presents a network to detect shadows by exploring and combining global context in deep layers and local context in shallow layers of a deep convolutional neural network (CNN). There are two technical contributions in our network design. First, we formulate the recurrent attention residual (RAR) module to combine the contexts in two adjacent CNN layers and learn an attention map to select a residual and then refine the context features. Second, we develop a bidirectional feature pyramid network (BFPN) to aggregate shadow contexts spanned across different CNN layers by deploying two series of RAR modules in the network to iteratively combine and refine context features: one series to refine context features from deep to shallow layers, and another series from shallow to deep layers. Hence, we can better suppress false detections and enhance shadow details at the same time. We evaluate our network on two common shadow detection benchmark datasets: SBU and UCF. Experimental results show that our network outperforms the best existing method with 34.88% reduction on SBU and 34.57% reduction on UCF for the balance error rate.
Tasks Shadow Detection
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/Lei_Zhu_Bi-directional_Feature_Pyramid_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/Lei_Zhu_Bi-directional_Feature_Pyramid_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/bidirectional-feature-pyramid-network-with
Repo https://github.com/zijundeng/BDRAR
Framework pytorch

RNN Simulations of Grammaticality Judgments on Long-distance Dependencies

Title RNN Simulations of Grammaticality Judgments on Long-distance Dependencies
Authors Shammur Absar Chowdhury, Roberto Zamparelli
Abstract The paper explores the ability of LSTM networks trained on a language modeling task to detect linguistic structures which are ungrammatical due to extraction violations (extra arguments and subject-relative clause island violations), and considers its implications for the debate on language innatism. The results show that the current RNN model can correctly classify (un)grammatical sentences, in certain conditions, but it is sensitive to linguistic processing factors and probably ultimately unable to induce a more abstract notion of grammaticality, at least in the domain we tested.
Tasks Language Modelling
Published 2018-08-01
URL https://www.aclweb.org/anthology/C18-1012/
PDF https://www.aclweb.org/anthology/C18-1012
PWC https://paperswithcode.com/paper/rnn-simulations-of-grammaticality-judgments
Repo https://github.com/LiCo-TREiL/Computational-Ungrammaticality
Framework none

Extracting Relational Facts by an End-to-End Neural Model with Copy Mechanism

Title Extracting Relational Facts by an End-to-End Neural Model with Copy Mechanism
Authors Xiangrong Zeng, Daojian Zeng, Shizhu He, Kang Liu, Jun Zhao
Abstract The relational facts in sentences are often complicated. Different relational triplets may have overlaps in a sentence. We divided the sentences into three types according to triplet overlap degree, including Normal, EntityPairOverlap and SingleEntiyOverlap. Existing methods mainly focus on Normal class and fail to extract relational triplets precisely. In this paper, we propose an end-to-end model based on sequence-to-sequence learning with copy mechanism, which can jointly extract relational facts from sentences of any of these classes. We adopt two different strategies in decoding process: employing only one united decoder or applying multiple separated decoders. We test our models in two public datasets and our model outperform the baseline method significantly.
Tasks Feature Engineering, Relation Extraction
Published 2018-07-01
URL https://www.aclweb.org/anthology/P18-1047/
PDF https://www.aclweb.org/anthology/P18-1047
PWC https://paperswithcode.com/paper/extracting-relational-facts-by-an-end-to-end
Repo https://github.com/xiangrongzeng/copy_re
Framework none

Experiments with Convolutional Neural Networks for Multi-Label Authorship Attribution

Title Experiments with Convolutional Neural Networks for Multi-Label Authorship Attribution
Authors Dainis Boumber, Yifan Zhang, Arjun Mukherjee
Abstract
Tasks Domain Adaptation
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1409/
PDF https://www.aclweb.org/anthology/L18-1409
PWC https://paperswithcode.com/paper/experiments-with-convolutional-neural
Repo https://github.com/dainis-boumber/AA_CNN
Framework tf

Deep Learning for Hand Gesture Recognition on Skeletal Data

Title Deep Learning for Hand Gesture Recognition on Skeletal Data
Authors Guillaume Devineau, Wang Xi, Jie Yang, Fabien Moutarde
Abstract In this paper, we introduce a new 3D hand gesture recognition approach based on a deep learning model. We introduce a new Convolutional Neural Network (CNN) where sequences of hand-skeletal joints’ positions are processed by parallel convolutions; we then investigate the performance of this model on hand gesture sequence classification tasks. Our model only uses hand-skeletal data and no depth image. Experimental results show that our approach achieves a state-of-the-art performance on a challenging dataset (DHG dataset from the SHREC 2017 3D Shape Retrieval Contest), when compared to other published approaches. Our model achieves a 91.28% classification accuracy for the 14 gesture classes case and an 84.35% classification accuracy for the 28 gesture classes case.
Tasks 3D Shape Retrieval, Gesture Recognition, Hand Gesture Recognition, Hand-Gesture Recognition, Skeleton Based Action Recognition, Temporal Information Extraction
Published 2018-05-15
URL https://hal-mines-paristech.archives-ouvertes.fr/hal-01737771/file/DeepLearning-HandSkeletalGestureRecognition_MINES-ParisTech_FG2018.pdf
PDF https://hal-mines-paristech.archives-ouvertes.fr/hal-01737771/file/DeepLearning-HandSkeletalGestureRecognition_MINES-ParisTech_FG2018.pdf
PWC https://paperswithcode.com/paper/deep-learning-for-hand-gesture-recognition-on
Repo https://github.com/guillaumephd/deep_learning_hand_gesture_recognition
Framework pytorch

Construction of Large-scale English Verbal Multiword Expression Annotated Corpus

Title Construction of Large-scale English Verbal Multiword Expression Annotated Corpus
Authors Akihiko Kato, Hiroyuki Shindo, Yuji Matsumoto
Abstract
Tasks Dependency Parsing, Semantic Parsing
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1396/
PDF https://www.aclweb.org/anthology/L18-1396
PWC https://paperswithcode.com/paper/construction-of-large-scale-english-verbal
Repo https://github.com/naist-cl-parsing/Verbal-MWE-annotations
Framework none
comments powered by Disqus