Paper Group AWR 41
SWRL2SPIN: A tool for transforming SWRL rule bases in OWL ontologies to object-oriented SPIN rules. Spoken SQuAD: A Study of Mitigating the Impact of Speech Recognition Errors on Listening Comprehension. A Fast Proximal Point Method for Computing Exact Wasserstein Distance. Graph Convolutional Networks for Text Classification. Efficient Neural Netw …
SWRL2SPIN: A tool for transforming SWRL rule bases in OWL ontologies to object-oriented SPIN rules
Title | SWRL2SPIN: A tool for transforming SWRL rule bases in OWL ontologies to object-oriented SPIN rules |
Authors | Nick Bassiliades |
Abstract | Semantic Web Rule Language (SWRL) combines OWL (Web Ontology Language) ontologies with Horn Logic rules of the Rule Markup Language (RuleML) family. Being supported by ontology editors, rule engines and ontology reasoners, it has become a very popular choice for developing rule-based applications on top of ontologies. However, SWRL is probably not go-ing to become a WWW Consortium standard, prohibiting industrial acceptance. On the other hand, SPIN (SPARQL Inferencing Notation) has become a de-facto industry standard to rep-resent SPARQL rules and constraints on Semantic Web models, building on the widespread acceptance of SPARQL (SPARQL Protocol and RDF Query Language). In this paper, we ar-gue that the life of existing SWRL rule-based ontology applications can be prolonged by con-verting them to SPIN. To this end, we have developed the SWRL2SPIN tool in Prolog that transforms SWRL rules into SPIN rules, considering the object-orientation of SPIN, i.e. linking rules to the appropriate ontology classes and optimizing them, as derived by analysing the rule conditions. |
Tasks | |
Published | 2018-01-27 |
URL | http://arxiv.org/abs/1801.09061v3 |
http://arxiv.org/pdf/1801.09061v3.pdf | |
PWC | https://paperswithcode.com/paper/swrl2spin-a-tool-for-transforming-swrl-rule |
Repo | https://github.com/nbassili/SWRL2SPIN |
Framework | none |
Spoken SQuAD: A Study of Mitigating the Impact of Speech Recognition Errors on Listening Comprehension
Title | Spoken SQuAD: A Study of Mitigating the Impact of Speech Recognition Errors on Listening Comprehension |
Authors | Chia-Hsuan Li, Szu-Lin Wu, Chi-Liang Liu, Hung-yi Lee |
Abstract | Reading comprehension has been widely studied. One of the most representative reading comprehension tasks is Stanford Question Answering Dataset (SQuAD), on which machine is already comparable with human. On the other hand, accessing large collections of multimedia or spoken content is much more difficult and time-consuming than plain text content for humans. It’s therefore highly attractive to develop machines which can automatically understand spoken content. In this paper, we propose a new listening comprehension task - Spoken SQuAD. On the new task, we found that speech recognition errors have catastrophic impact on machine comprehension, and several approaches are proposed to mitigate the impact. |
Tasks | Question Answering, Reading Comprehension, Speech Recognition |
Published | 2018-04-01 |
URL | http://arxiv.org/abs/1804.00320v1 |
http://arxiv.org/pdf/1804.00320v1.pdf | |
PWC | https://paperswithcode.com/paper/spoken-squad-a-study-of-mitigating-the-impact |
Repo | https://github.com/chiahsuan156/Spoken-SQuAD |
Framework | none |
A Fast Proximal Point Method for Computing Exact Wasserstein Distance
Title | A Fast Proximal Point Method for Computing Exact Wasserstein Distance |
Authors | Yujia Xie, Xiangfeng Wang, Ruijia Wang, Hongyuan Zha |
Abstract | Wasserstein distance plays increasingly important roles in machine learning, stochastic programming and image processing. Major efforts have been under way to address its high computational complexity, some leading to approximate or regularized variations such as Sinkhorn distance. However, as we will demonstrate, regularized variations with large regularization parameter will degradate the performance in several important machine learning applications, and small regularization parameter will fail due to numerical stability issues with existing algorithms. We address this challenge by developing an Inexact Proximal point method for exact Optimal Transport problem (IPOT) with the proximal operator approximately evaluated at each iteration using projections to the probability simplex. The algorithm (a) converges to exact Wasserstein distance with theoretical guarantee and robust regularization parameter selection, (b) alleviates numerical stability issue, (c) has similar computational complexity to Sinkhorn, and (d) avoids the shrinking problem when apply to generative models. Furthermore, a new algorithm is proposed based on IPOT to obtain sharper Wasserstein barycenter. |
Tasks | |
Published | 2018-02-12 |
URL | https://arxiv.org/abs/1802.04307v3 |
https://arxiv.org/pdf/1802.04307v3.pdf | |
PWC | https://paperswithcode.com/paper/a-fast-proximal-point-method-for-computing |
Repo | https://github.com/xieyujia/IPOT |
Framework | tf |
Graph Convolutional Networks for Text Classification
Title | Graph Convolutional Networks for Text Classification |
Authors | Liang Yao, Chengsheng Mao, Yuan Luo |
Abstract | Text classification is an important and classical problem in natural language processing. There have been a number of studies that applied convolutional neural networks (convolution on regular grid, e.g., sequence) to classification. However, only a limited number of studies have explored the more flexible graph convolutional neural networks (convolution on non-grid, e.g., arbitrary graph) for the task. In this work, we propose to use graph convolutional networks for text classification. We build a single text graph for a corpus based on word co-occurrence and document word relations, then learn a Text Graph Convolutional Network (Text GCN) for the corpus. Our Text GCN is initialized with one-hot representation for word and document, it then jointly learns the embeddings for both words and documents, as supervised by the known class labels for documents. Our experimental results on multiple benchmark datasets demonstrate that a vanilla Text GCN without any external word embeddings or knowledge outperforms state-of-the-art methods for text classification. On the other hand, Text GCN also learns predictive word and document embeddings. In addition, experimental results show that the improvement of Text GCN over state-of-the-art comparison methods become more prominent as we lower the percentage of training data, suggesting the robustness of Text GCN to less training data in text classification. |
Tasks | Sentiment Analysis, Text Classification |
Published | 2018-09-15 |
URL | http://arxiv.org/abs/1809.05679v3 |
http://arxiv.org/pdf/1809.05679v3.pdf | |
PWC | https://paperswithcode.com/paper/graph-convolutional-networks-for-text |
Repo | https://github.com/plkmo/Bible_Text_GCN |
Framework | pytorch |
Efficient Neural Network Robustness Certification with General Activation Functions
Title | Efficient Neural Network Robustness Certification with General Activation Functions |
Authors | Huan Zhang, Tsui-Wei Weng, Pin-Yu Chen, Cho-Jui Hsieh, Luca Daniel |
Abstract | Finding minimum distortion of adversarial examples and thus certifying robustness in neural network classifiers for given data points is known to be a challenging problem. Nevertheless, recently it has been shown to be possible to give a non-trivial certified lower bound of minimum adversarial distortion, and some recent progress has been made towards this direction by exploiting the piece-wise linear nature of ReLU activations. However, a generic robustness certification for general activation functions still remains largely unexplored. To address this issue, in this paper we introduce CROWN, a general framework to certify robustness of neural networks with general activation functions for given input data points. The novelty in our algorithm consists of bounding a given activation function with linear and quadratic functions, hence allowing it to tackle general activation functions including but not limited to four popular choices: ReLU, tanh, sigmoid and arctan. In addition, we facilitate the search for a tighter certified lower bound by adaptively selecting appropriate surrogates for each neuron activation. Experimental results show that CROWN on ReLU networks can notably improve the certified lower bounds compared to the current state-of-the-art algorithm Fast-Lin, while having comparable computational efficiency. Furthermore, CROWN also demonstrates its effectiveness and flexibility on networks with general activation functions, including tanh, sigmoid and arctan. |
Tasks | |
Published | 2018-11-02 |
URL | http://arxiv.org/abs/1811.00866v1 |
http://arxiv.org/pdf/1811.00866v1.pdf | |
PWC | https://paperswithcode.com/paper/efficient-neural-network-robustness |
Repo | https://github.com/huanzhang12/RecurJac-and-CROWN |
Framework | tf |
Neural Rejuvenation: Improving Deep Network Training by Enhancing Computational Resource Utilization
Title | Neural Rejuvenation: Improving Deep Network Training by Enhancing Computational Resource Utilization |
Authors | Siyuan Qiao, Zhe Lin, Jianming Zhang, Alan Yuille |
Abstract | In this paper, we study the problem of improving computational resource utilization of neural networks. Deep neural networks are usually over-parameterized for their tasks in order to achieve good performances, thus are likely to have underutilized computational resources. This observation motivates a lot of research topics, e.g. network pruning, architecture search, etc. As models with higher computational costs (e.g. more parameters or more computations) usually have better performances, we study the problem of improving the resource utilization of neural networks so that their potentials can be further realized. To this end, we propose a novel optimization method named Neural Rejuvenation. As its name suggests, our method detects dead neurons and computes resource utilization in real time, rejuvenates dead neurons by resource reallocation and reinitialization, and trains them with new training schemes. By simply replacing standard optimizers with Neural Rejuvenation, we are able to improve the performances of neural networks by a very large margin while using similar training efforts and maintaining their original resource usages. |
Tasks | Network Pruning, Neural Architecture Search |
Published | 2018-12-02 |
URL | http://arxiv.org/abs/1812.00481v1 |
http://arxiv.org/pdf/1812.00481v1.pdf | |
PWC | https://paperswithcode.com/paper/neural-rejuvenation-improving-deep-network |
Repo | https://github.com/joe-siyuan-qiao/NeuralRejuvenation-CVPR19 |
Framework | pytorch |
On the End-to-End Solution to Mandarin-English Code-switching Speech Recognition
Title | On the End-to-End Solution to Mandarin-English Code-switching Speech Recognition |
Authors | Zhiping Zeng, Yerbolat Khassanov, Van Tung Pham, Haihua Xu, Eng Siong Chng, Haizhou Li |
Abstract | Code-switching (CS) refers to a linguistic phenomenon where a speaker uses different languages in an utterance or between alternating utterances. In this work, we study end-to-end (E2E) approaches to the Mandarin-English code-switching speech recognition (CSSR) task. We first examine the effectiveness of using data augmentation and byte-pair encoding (BPE) subword units. More importantly, we propose a multitask learning recipe, where a language identification task is explicitly learned in addition to the E2E speech recognition task. Furthermore, we introduce an efficient word vocabulary expansion method for language modeling to alleviate data sparsity issues under the code-switching scenario. Experimental results on the SEAME data, a Mandarin-English CS corpus, demonstrate the effectiveness of the proposed methods. |
Tasks | Data Augmentation, Language Identification, Language Modelling, Speech Recognition |
Published | 2018-11-01 |
URL | https://arxiv.org/abs/1811.00241v2 |
https://arxiv.org/pdf/1811.00241v2.pdf | |
PWC | https://paperswithcode.com/paper/on-the-end-to-end-solution-to-mandarin |
Repo | https://github.com/zengzp0912/SEAME-dev-set |
Framework | none |
Nose, eyes and ears: Head pose estimation by locating facial keypoints
Title | Nose, eyes and ears: Head pose estimation by locating facial keypoints |
Authors | Aryaman Gupta, Kalpit Thakkar, Vineet Gandhi, P J Narayanan |
Abstract | Monocular head pose estimation requires learning a model that computes the intrinsic Euler angles for pose (yaw, pitch, roll) from an input image of human face. Annotating ground truth head pose angles for images in the wild is difficult and requires ad-hoc fitting procedures (which provides only coarse and approximate annotations). This highlights the need for approaches which can train on data captured in controlled environment and generalize on the images in the wild (with varying appearance and illumination of the face). Most present day deep learning approaches which learn a regression function directly on the input images fail to do so. To this end, we propose to use a higher level representation to regress the head pose while using deep learning architectures. More specifically, we use the uncertainty maps in the form of 2D soft localization heatmap images over five facial keypoints, namely left ear, right ear, left eye, right eye and nose, and pass them through an convolutional neural network to regress the head-pose. We show head pose estimation results on two challenging benchmarks BIWI and AFLW and our approach surpasses the state of the art on both the datasets. |
Tasks | Head Pose Estimation, Pose Estimation |
Published | 2018-12-03 |
URL | http://arxiv.org/abs/1812.00739v1 |
http://arxiv.org/pdf/1812.00739v1.pdf | |
PWC | https://paperswithcode.com/paper/nose-eyes-and-ears-head-pose-estimation-by |
Repo | https://github.com/manoj901/HeadposeEstimation |
Framework | pytorch |
Learning Rich Features for Image Manipulation Detection
Title | Learning Rich Features for Image Manipulation Detection |
Authors | Peng Zhou, Xintong Han, Vlad I. Morariu, Larry S. Davis |
Abstract | Image manipulation detection is different from traditional semantic object detection because it pays more attention to tampering artifacts than to image content, which suggests that richer features need to be learned. We propose a two-stream Faster R-CNN network and train it endto- end to detect the tampered regions given a manipulated image. One of the two streams is an RGB stream whose purpose is to extract features from the RGB image input to find tampering artifacts like strong contrast difference, unnatural tampered boundaries, and so on. The other is a noise stream that leverages the noise features extracted from a steganalysis rich model filter layer to discover the noise inconsistency between authentic and tampered regions. We then fuse features from the two streams through a bilinear pooling layer to further incorporate spatial co-occurrence of these two modalities. Experiments on four standard image manipulation datasets demonstrate that our two-stream framework outperforms each individual stream, and also achieves state-of-the-art performance compared to alternative methods with robustness to resizing and compression. |
Tasks | Image Manipulation Detection, Object Detection |
Published | 2018-05-13 |
URL | http://arxiv.org/abs/1805.04953v1 |
http://arxiv.org/pdf/1805.04953v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-rich-features-for-image-manipulation |
Repo | https://github.com/pengzhou1108/RGB-N |
Framework | tf |
Neural Abstractive Text Summarization with Sequence-to-Sequence Models: A Survey
Title | Neural Abstractive Text Summarization with Sequence-to-Sequence Models: A Survey |
Authors | Tian Shi, Yaser Keneshloo, Naren Ramakrishnan, Chandan K. Reddy |
Abstract | In the past few years, neural abstractive text summarization with sequence-to-sequence (seq2seq) models have gained a lot of popularity. Many interesting techniques have been proposed to improve the seq2seq models, making them capable of handling different challenges, such as saliency, fluency and human readability, and generate high-quality summaries. Generally speaking, most of these techniques differ in one of these three categories: network structure, parameter inference, and decoding/generation. There are also other concerns, such as efficiency and parallelism for training a model. In this paper, we provide a comprehensive literature and technical survey on different seq2seq models for abstractive text summarization from viewpoint of network structures, training strategies, and summary generation algorithms. Many models were first proposed for language modeling and generation tasks, such as machine translation, and later applied to abstractive text summarization. Therefore, we also provide a brief review of these models. As part of this survey, we also develop an open source library, namely Neural Abstractive Text Summarizer (NATS) toolkit, for the abstractive text summarization. An extensive set of experiments have been conducted on the widely used CNN/Daily Mail dataset to examine the effectiveness of several different neural network components. Finally, we benchmark two models implemented in NATS on two recently released datasets, i.e., Newsroom and Bytecup. |
Tasks | Abstractive Text Summarization, Language Modelling, Machine Translation, Text Summarization |
Published | 2018-12-05 |
URL | https://arxiv.org/abs/1812.02303v3 |
https://arxiv.org/pdf/1812.02303v3.pdf | |
PWC | https://paperswithcode.com/paper/neural-abstractive-text-summarization-with |
Repo | https://github.com/sf-18/politicalsynthesis |
Framework | tf |
Discourse Coherence in the Wild: A Dataset, Evaluation and Methods
Title | Discourse Coherence in the Wild: A Dataset, Evaluation and Methods |
Authors | Alice Lai, Joel Tetreault |
Abstract | To date there has been very little work on assessing discourse coherence methods on real-world data. To address this, we present a new corpus of real-world texts (GCDC) as well as the first large-scale evaluation of leading discourse coherence algorithms. We show that neural models, including two that we introduce here (SentAvg and ParSeq), tend to perform best. We analyze these performance differences and discuss patterns we observed in low coherence texts in four domains. |
Tasks | |
Published | 2018-05-14 |
URL | http://arxiv.org/abs/1805.04993v1 |
http://arxiv.org/pdf/1805.04993v1.pdf | |
PWC | https://paperswithcode.com/paper/discourse-coherence-in-the-wild-a-dataset |
Repo | https://github.com/aylai/GCDC-corpus |
Framework | none |
Deep Neural Networks for Bot Detection
Title | Deep Neural Networks for Bot Detection |
Authors | Sneha Kudugunta, Emilio Ferrara |
Abstract | The problem of detecting bots, automated social media accounts governed by software but disguising as human users, has strong implications. For example, bots have been used to sway political elections by distorting online discourse, to manipulate the stock market, or to push anti-vaccine conspiracy theories that caused health epidemics. Most techniques proposed to date detect bots at the account level, by processing large amount of social media posts, and leveraging information from network structure, temporal dynamics, sentiment analysis, etc. In this paper, we propose a deep neural network based on contextual long short-term memory (LSTM) architecture that exploits both content and metadata to detect bots at the tweet level: contextual features are extracted from user metadata and fed as auxiliary input to LSTM deep nets processing the tweet text. Another contribution that we make is proposing a technique based on synthetic minority oversampling to generate a large labeled dataset, suitable for deep nets training, from a minimal amount of labeled data (roughly 3,000 examples of sophisticated Twitter bots). We demonstrate that, from just one single tweet, our architecture can achieve high classification accuracy (AUC > 96%) in separating bots from humans. We apply the same architecture to account-level bot detection, achieving nearly perfect classification accuracy (AUC > 99%). Our system outperforms previous state of the art while leveraging a small and interpretable set of features yet requiring minimal training data. |
Tasks | Sentiment Analysis |
Published | 2018-02-12 |
URL | http://arxiv.org/abs/1802.04289v2 |
http://arxiv.org/pdf/1802.04289v2.pdf | |
PWC | https://paperswithcode.com/paper/deep-neural-networks-for-bot-detection |
Repo | https://github.com/ji10bhatt/Capstone-Social-Media-Bot-Detection |
Framework | none |
Memory In Memory: A Predictive Neural Network for Learning Higher-Order Non-Stationarity from Spatiotemporal Dynamics
Title | Memory In Memory: A Predictive Neural Network for Learning Higher-Order Non-Stationarity from Spatiotemporal Dynamics |
Authors | Yunbo Wang, Jianjin Zhang, Hongyu Zhu, Mingsheng Long, Jianmin Wang, Philip S Yu |
Abstract | Natural spatiotemporal processes can be highly non-stationary in many ways, e.g. the low-level non-stationarity such as spatial correlations or temporal dependencies of local pixel values; and the high-level variations such as the accumulation, deformation or dissipation of radar echoes in precipitation forecasting. From Cramer’s Decomposition, any non-stationary process can be decomposed into deterministic, time-variant polynomials, plus a zero-mean stochastic term. By applying differencing operations appropriately, we may turn time-variant polynomials into a constant, making the deterministic component predictable. However, most previous recurrent neural networks for spatiotemporal prediction do not use the differential signals effectively, and their relatively simple state transition functions prevent them from learning too complicated variations in spacetime. We propose the Memory In Memory (MIM) networks and corresponding recurrent blocks for this purpose. The MIM blocks exploit the differential signals between adjacent recurrent states to model the non-stationary and approximately stationary properties in spatiotemporal dynamics with two cascaded, self-renewed memory modules. By stacking multiple MIM blocks, we could potentially handle higher-order non-stationarity. The MIM networks achieve the state-of-the-art results on four spatiotemporal prediction tasks across both synthetic and real-world datasets. We believe that the general idea of this work can be potentially applied to other time-series forecasting tasks. |
Tasks | Time Series, Time Series Forecasting, Video Prediction |
Published | 2018-11-19 |
URL | http://arxiv.org/abs/1811.07490v3 |
http://arxiv.org/pdf/1811.07490v3.pdf | |
PWC | https://paperswithcode.com/paper/memory-in-memory-a-predictive-neural-network |
Repo | https://github.com/Yunbo426/MIM |
Framework | tf |
Multi-Level Network Embedding with Boosted Low-Rank Matrix Approximation
Title | Multi-Level Network Embedding with Boosted Low-Rank Matrix Approximation |
Authors | Jundong Li, Liang Wu, Huan Liu |
Abstract | As opposed to manual feature engineering which is tedious and difficult to scale, network representation learning has attracted a surge of research interests as it automates the process of feature learning on graphs. The learned low-dimensional node vector representation is generalizable and eases the knowledge discovery process on graphs by enabling various off-the-shelf machine learning tools to be directly applied. Recent research has shown that the past decade of network embedding approaches either explicitly factorize a carefully designed matrix to obtain the low-dimensional node vector representation or are closely related to implicit matrix factorization, with the fundamental assumption that the factorized node connectivity matrix is low-rank. Nonetheless, the global low-rank assumption does not necessarily hold especially when the factorized matrix encodes complex node interactions, and the resultant single low-rank embedding matrix is insufficient to capture all the observed connectivity patterns. In this regard, we propose a novel multi-level network embedding framework BoostNE, which can learn multiple network embedding representations of different granularity from coarse to fine without imposing the prevalent global low-rank assumption. The proposed BoostNE method is also in line with the successful gradient boosting method in ensemble learning as multiple weak embeddings lead to a stronger and more effective one. We assess the effectiveness of the proposed BoostNE framework by comparing it with existing state-of-the-art network embedding methods on various datasets, and the experimental results corroborate the superiority of the proposed BoostNE network embedding framework. |
Tasks | Feature Engineering, Network Embedding, Representation Learning |
Published | 2018-08-26 |
URL | http://arxiv.org/abs/1808.08627v1 |
http://arxiv.org/pdf/1808.08627v1.pdf | |
PWC | https://paperswithcode.com/paper/multi-level-network-embedding-with-boosted |
Repo | https://github.com/benedekrozemberczki/karateclub |
Framework | none |
A multi-instance deep neural network classifier: application to Higgs boson CP measurement
Title | A multi-instance deep neural network classifier: application to Higgs boson CP measurement |
Authors | P. Bialas, D. Nemeth, E. Richter-Wąs |
Abstract | We investigate properties of a classifier applied to the measurements of the CP state of the Higgs boson in $H\rightarrow\tau\tau$ decays. The problem is framed as binary classifier applied to individual instances. Then the prior knowledge that the instances belong to the same class is used to define the multi-instance classifier. Its final score is calculated as multiplication of single instance scores for a given series of instances. In the paper we discuss properties of such classifier, notably its dependence on the number of instances in the series. This classifier exhibits very strong random dependence on the number of epochs used for training and requires careful tuning of the classification threshold. We derive formula for this optimal threshold. |
Tasks | |
Published | 2018-03-02 |
URL | http://arxiv.org/abs/1803.00838v2 |
http://arxiv.org/pdf/1803.00838v2.pdf | |
PWC | https://paperswithcode.com/paper/a-multi-instance-deep-neural-network |
Repo | https://github.com/klasocha/HiggsCP |
Framework | none |