January 24, 2020

2508 words 12 mins read

Paper Group NANR 114

Paper Group NANR 114

JPEG Artifacts Reduction via Deep Convolutional Sparse Coding. OSIAN: Open Source International Arabic News Corpus - Preparation and Integration into the CLARIN-infrastructure. An Automated Framework for Fast Cognate Detection and Bayesian Phylogenetic Inference in Computational Historical Linguistics. A NON-LINEAR THEORY FOR SENTENCE EMBEDDING. Te …

JPEG Artifacts Reduction via Deep Convolutional Sparse Coding

Title JPEG Artifacts Reduction via Deep Convolutional Sparse Coding
Authors Xueyang Fu, Zheng-Jun Zha, Feng Wu, Xinghao Ding, John Paisley
Abstract To effectively reduce JPEG compression artifacts, we propose a deep convolutional sparse coding (DCSC) network architecture. We design our DCSC in the framework of classic learned iterative shrinkage-threshold algorithm. To focus on recognizing and separating artifacts only, we sparsely code the feature maps instead of the raw image. The final de-blocked image is directly reconstructed from the coded features. We use dilated convolution to extract multi-scale image features, which allows our single model to simultaneously handle multiple JPEG compression levels. Since our method integrates model-based convolutional sparse coding with a learning-based deep neural network, the entire network structure is compact and more explainable. The resulting lightweight model generates comparable or better de-blocking results when compared with state-of-the-art methods.
Tasks
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Fu_JPEG_Artifacts_Reduction_via_Deep_Convolutional_Sparse_Coding_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Fu_JPEG_Artifacts_Reduction_via_Deep_Convolutional_Sparse_Coding_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/jpeg-artifacts-reduction-via-deep
Repo
Framework

OSIAN: Open Source International Arabic News Corpus - Preparation and Integration into the CLARIN-infrastructure

Title OSIAN: Open Source International Arabic News Corpus - Preparation and Integration into the CLARIN-infrastructure
Authors Imad Zeroual, Dirk Goldhahn, Thomas Eckart, Abdelhak Lakhouaja
Abstract The World Wide Web has become a fundamental resource for building large text corpora. Broadcasting platforms such as news websites are rich sources of data regarding diverse topics and form a valuable foundation for research. The Arabic language is extensively utilized on the Web. Still, Arabic is relatively an under-resourced language in terms of availability of freely annotated corpora. This paper presents the first version of the Open Source International Arabic News (OSIAN) corpus. The corpus data was collected from international Arabic news websites, all being freely available on the Web. The corpus consists of about 3.5 million articles comprising more than 37 million sentences and roughly 1 billion tokens. It is encoded in XML; each article is annotated with metadata information. Moreover, each word is annotated with lemma and part-of-speech. the described corpus is processed, archived and published into the CLARIN infrastructure. This publication includes descriptive metadata via OAI-PMH, direct access to the plain text material (available under Creative Commons Attribution-Non-Commercial 4.0 International License - CC BY-NC 4.0), and integration into the WebLicht annotation platform and CLARIN{'}s Federated Content Search FCS.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4619/
PDF https://www.aclweb.org/anthology/W19-4619
PWC https://paperswithcode.com/paper/osian-open-source-international-arabic-news
Repo
Framework

An Automated Framework for Fast Cognate Detection and Bayesian Phylogenetic Inference in Computational Historical Linguistics

Title An Automated Framework for Fast Cognate Detection and Bayesian Phylogenetic Inference in Computational Historical Linguistics
Authors Taraka Rama, Johann-Mattis List
Abstract We present a fully automated workflow for phylogenetic reconstruction on large datasets, consisting of two novel methods, one for fast detection of cognates and one for fast Bayesian phylogenetic inference. Our results show that the methods take less than a few minutes to process language families that have so far required large amounts of time and computational power. Moreover, the cognates and the trees inferred from the method are quite close, both to gold standard cognate judgments and to expert language family trees. Given its speed and ease of application, our framework is specifically useful for the exploration of very large datasets in historical linguistics.
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1627/
PDF https://www.aclweb.org/anthology/P19-1627
PWC https://paperswithcode.com/paper/an-automated-framework-for-fast-cognate
Repo
Framework

A NON-LINEAR THEORY FOR SENTENCE EMBEDDING

Title A NON-LINEAR THEORY FOR SENTENCE EMBEDDING
Authors Hichem Mezaoui, Isar Nejadgholi
Abstract This paper revisits the Random Walk model for sentence embedding in the context of non-extensive statistics. We propose a non-extensive algebra to compute the discourse vector. We argue that by doing so we are taking into account high non-linearity in the semantic space. Furthermore, we show that by considering a non-extensive algebra, the compounding effect of the vector length is mitigated. Overall, we show that the proposed model leads to good sentence embedding. We evaluate the embedding method on textual similarity tasks.
Tasks Sentence Embedding
Published 2019-05-01
URL https://openreview.net/forum?id=SJMZRsC9Y7
PDF https://openreview.net/pdf?id=SJMZRsC9Y7
PWC https://paperswithcode.com/paper/a-non-linear-theory-for-sentence-embedding
Repo
Framework

Team Taurus at SemEval-2019 Task 9: Expert-informed pattern recognition for suggestion mining

Title Team Taurus at SemEval-2019 Task 9: Expert-informed pattern recognition for suggestion mining
Authors Nelleke Oostdijk, Hans van Halteren
Abstract This paper presents our submissions to SemEval-2019 Task9, Suggestion Mining. Our system is one in a series of systems in which we compare an approach using expert-defined rules with a comparable one using machine learning. We target tasks with a syntactic or semantic component that might be better described by a human understanding the task than by a machine learner only able to count features. For Semeval-2019 Task 9, the expert rules clearly outperformed our machine learning model when training and testing on equally balanced testsets.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2219/
PDF https://www.aclweb.org/anthology/S19-2219
PWC https://paperswithcode.com/paper/team-taurus-at-semeval-2019-task-9-expert
Repo
Framework

Multi-Task, Multi-Channel, Multi-Input Learning for Mental Illness Detection using Social Media Text

Title Multi-Task, Multi-Channel, Multi-Input Learning for Mental Illness Detection using Social Media Text
Authors Prasadith Kirinde Gamaarachchige, Diana Inkpen
Abstract We investigate the impact of using emotional patterns identified by the clinical practitioners and computational linguists to enhance the prediction capabilities of a mental illness detection (in our case depression and post-traumatic stress disorder) model built using a deep neural network architecture. Over the years, deep learning methods have been successfully used in natural language processing tasks, including a few in the domain of mental illness and suicide ideation detection. We illustrate the effectiveness of using multi-task learning with a multi-channel convolutional neural network as the shared representation and use additional inputs identified by researchers as indicatives in detecting mental disorders to enhance the model predictability. Given the limited amount of unstructured data available for training, we managed to obtain a task-specific AUC higher than 0.90. In comparison to methods such as multi-class classification, we identified multi-task learning with multi-channel convolution neural network and multiple-inputs to be effective in detecting mental disorders.
Tasks Multi-Task Learning
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-6208/
PDF https://www.aclweb.org/anthology/D19-6208
PWC https://paperswithcode.com/paper/multi-task-multi-channel-multi-input-learning
Repo
Framework

Neural Multi-Task Learning for Stance Prediction

Title Neural Multi-Task Learning for Stance Prediction
Authors Wei Fang, Moin Nadeem, Mitra Mohtarami, James Glass
Abstract We present a multi-task learning model that leverages large amount of textual information from existing datasets to improve stance prediction. In particular, we utilize multiple NLP tasks under both unsupervised and supervised settings for the target stance prediction task. Our model obtains state-of-the-art performance on a public benchmark dataset, Fake News Challenge, outperforming current approaches by a wide margin.
Tasks Multi-Task Learning
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-6603/
PDF https://www.aclweb.org/anthology/D19-6603
PWC https://paperswithcode.com/paper/neural-multi-task-learning-for-stance
Repo
Framework

ChiMed: A Chinese Medical Corpus for Question Answering

Title ChiMed: A Chinese Medical Corpus for Question Answering
Authors Yuanhe Tian, Weicheng Ma, Fei Xia, Yan Song
Abstract Question answering (QA) is a challenging task in natural language processing (NLP), especially when it is applied to specific domains. While models trained in the general domain can be adapted to a new target domain, their performance often degrades significantly due to domain mismatch. Alternatively, one can require a large amount of domain-specific QA data, but such data are rare, especially for the medical domain. In this study, we first collect a large-scale Chinese medical QA corpus called ChiMed; second we annotate a small fraction of the corpus to check the quality of the answers; third, we extract two datasets from the corpus and use them for the relevancy prediction task and the adoption prediction task. Several benchmark models are applied to the datasets, producing good results for both tasks.
Tasks Question Answering
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5027/
PDF https://www.aclweb.org/anthology/W19-5027
PWC https://paperswithcode.com/paper/chimed-a-chinese-medical-corpus-for-question
Repo
Framework

Event Causality Recognition Exploiting Multiple Annotators’ Judgments and Background Knowledge

Title Event Causality Recognition Exploiting Multiple Annotators’ Judgments and Background Knowledge
Authors Kazuma Kadowaki, Ryu Iida, Kentaro Torisawa, Jong-Hoon Oh, Julien Kloetzer
Abstract We propose new BERT-based methods for recognizing event causality such as {}smoke cigarettes{''} {--}{\textgreater} {}die of lung cancer{''} written in web texts. In our methods, we grasp each annotator{'}s policy by training multiple classifiers, each of which predicts the labels given by a single annotator, and combine the resulting classifiers{'} outputs to predict the final labels determined by majority vote. Furthermore, we investigate the effect of supplying background knowledge to our classifiers. Since BERT models are pre-trained with a large corpus, some sort of background knowledge for event causality may be learned during pre-training. Our experiments with a Japanese dataset suggest that this is actually the case: Performance improved when we pre-trained the BERT models with web texts containing a large number of event causalities instead of Wikipedia articles or randomly sampled web texts. However, this effect was limited. Therefore, we further improved performance by simply adding texts related to an input causality candidate as background knowledge to the input of the BERT models. We believe these findings indicate a promising future research direction.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1590/
PDF https://www.aclweb.org/anthology/D19-1590
PWC https://paperswithcode.com/paper/event-causality-recognition-exploiting
Repo
Framework

Probabilistic Planning with Sequential Monte Carlo methods

Title Probabilistic Planning with Sequential Monte Carlo methods
Authors Alexandre Piche, Valentin Thomas, Cyril Ibrahim, Yoshua Bengio, Chris Pal
Abstract In this work, we propose a novel formulation of planning which views it as a probabilistic inference problem over future optimal trajectories. This enables us to use sampling methods, and thus, tackle planning in continuous domains using a fixed computational budget. We design a new algorithm, Sequential Monte Carlo Planning, by leveraging classical methods in Sequential Monte Carlo and Bayesian smoothing in the context of control as inference. Furthermore, we show that Sequential Monte Carlo Planning can capture multimodal policies and can quickly learn continuous control tasks.
Tasks Continuous Control
Published 2019-05-01
URL https://openreview.net/forum?id=ByetGn0cYX
PDF https://openreview.net/pdf?id=ByetGn0cYX
PWC https://paperswithcode.com/paper/probabilistic-planning-with-sequential-monte
Repo
Framework

Whitening and Coloring transform for GANs

Title Whitening and Coloring transform for GANs
Authors Aliaksandr Siarohin, Enver Sangineto, Nicu Sebe
Abstract Batch Normalization (BN) is a common technique used to speed-up and stabilize training. On the other hand, the learnable parameters of BN are commonly used in conditional Generative Adversarial Networks (cGANs) for representing class-specific information using conditional Batch Normalization (cBN). In this paper we propose to generalize both BN and cBN using a Whitening and Coloring based batch normalization. We show that our conditional Coloring can represent categorical conditioning information which largely helps the cGAN qualitative results. Moreover, we show that full-feature whitening is important in a general GAN scenario in which the training process is known to be highly unstable. We test our approach on different datasets and using different GAN networks and training protocols, showing a consistent improvement in all the tested frameworks. Our CIFAR-10 supervised results are higher than all previous works on this dataset.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=S1x2Fj0qKQ
PDF https://openreview.net/pdf?id=S1x2Fj0qKQ
PWC https://paperswithcode.com/paper/whitening-and-coloring-transform-for-gans
Repo
Framework

The LAIX Systems in the BEA-2019 GEC Shared Task

Title The LAIX Systems in the BEA-2019 GEC Shared Task
Authors Ruobing Li, Chuan Wang, Yefei Zha, Yonghong Yu, Shiman Guo, Qiang Wang, Yang Liu, Hui Lin
Abstract In this paper, we describe two systems we developed for the three tracks we have participated in the BEA-2019 GEC Shared Task. We investigate competitive classification models with bi-directional recurrent neural networks (Bi-RNN) and neural machine translation (NMT) models. For different tracks, we use ensemble systems to selectively combine the NMT models, the classification models, and some rules, and demonstrate that an ensemble solution can effectively improve GEC performance over single systems. Our GEC systems ranked the first in the Unrestricted Track, and the third in both the Restricted Track and the Low Resource Track.
Tasks Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4416/
PDF https://www.aclweb.org/anthology/W19-4416
PWC https://paperswithcode.com/paper/the-laix-systems-in-the-bea-2019-gec-shared
Repo
Framework

Tinkering with black boxes: counterfactuals uncover modularity in generative models

Title Tinkering with black boxes: counterfactuals uncover modularity in generative models
Authors Michel Besserve, Remy Sun, Bernhard Schoelkopf
Abstract Deep generative models such as Generative Adversarial Networks (GANs) and Variational Auto-Encoders (VAEs) are important tools to capture and investigate the properties of complex empirical data. However, the complexity of their inner elements makes their functionment challenging to assess and modify. In this respect, these architectures behave as black box models. In order to better understand the function of such networks, we analyze their modularity based on the counterfactual manipulation of their internal variables. Our experiments on the generation of human faces with VAEs and GANs support that modularity between activation maps distributed over channels of generator architectures is achieved to some degree, can be used to better understand how these systems operate and allow meaningful transformations of the generated images without further training. erate and edit the content of generated images.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=Byldr3RqKX
PDF https://openreview.net/pdf?id=Byldr3RqKX
PWC https://paperswithcode.com/paper/tinkering-with-black-boxes-counterfactuals
Repo
Framework

IM-Net for High Resolution Video Frame Interpolation

Title IM-Net for High Resolution Video Frame Interpolation
Authors Tomer Peleg, Pablo Szekely, Doron Sabo, Omry Sendik
Abstract Video frame interpolation is a long-studied problem in the video processing field. Recently, deep learning approaches have been applied to this problem, showing impressive results on low-resolution benchmarks. However, these methods do not scale-up favorably to high resolutions. Specifically, when the motion exceeds a typical number of pixels, their interpolation quality is degraded. Moreover, their run time renders them impractical for real-time applications. In this paper we propose IM-Net: an interpolated motion neural network. We use an economic structured architecture and end-to-end training with multi-scale tailored losses. In particular, we formulate interpolated motion estimation as classification rather than regression. IM-Net outperforms previous methods by more than 1.3dB (PSNR) on a high resolution version of the recently introduced Vimeo triplet dataset. Moreover, the network runs in less than 33msec on a single GPU for HD resolution.
Tasks Motion Estimation, Video Frame Interpolation
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Peleg_IM-Net_for_High_Resolution_Video_Frame_Interpolation_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Peleg_IM-Net_for_High_Resolution_Video_Frame_Interpolation_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/im-net-for-high-resolution-video-frame
Repo
Framework

Neural Network Regression with Beta, Dirichlet, and Dirichlet-Multinomial Outputs

Title Neural Network Regression with Beta, Dirichlet, and Dirichlet-Multinomial Outputs
Authors Peter Sadowski, Pierre Baldi
Abstract We propose a method for quantifying uncertainty in neural network regression models when the targets are real values on a $d$-dimensional simplex, such as probabilities. We show that each target can be modeled as a sample from a Dirichlet distribution, where the parameters of the Dirichlet are provided by the output of a neural network, and that the combined model can be trained using the gradient of the data likelihood. This approach provides interpretable predictions in the form of multidimensional distributions, rather than point estimates, from which one can obtain confidence intervals or quantify risk in decision making. Furthermore, we show that the same approach can be used to model targets in the form of empirical counts as samples from the Dirichlet-multinomial compound distribution. In experiments, we verify that our approach provides these benefits without harming the performance of the point estimate predictions on two diverse applications: (1) distilling deep convolutional networks trained on CIFAR-100, and (2) predicting the location of particle collisions in the XENON1T Dark Matter detector.
Tasks Decision Making
Published 2019-05-01
URL https://openreview.net/forum?id=BJeRg205Fm
PDF https://openreview.net/pdf?id=BJeRg205Fm
PWC https://paperswithcode.com/paper/neural-network-regression-with-beta-dirichlet
Repo
Framework
comments powered by Disqus