October 16, 2019

2789 words 14 mins read

Paper Group NAWR 21

Paper Group NAWR 21

Multi-Task Learning for Document Ranking and Query Suggestion. Image Deblurring with a Class-Specific Prior. Identifying Key Sentences for Precision Oncology Using Semi-Supervised Learning. Introducing a Lexicon of Verbal Polarity Shifters for English. Fast Adaptive RNN Encoder-Decoder for Anomaly Detection in SMD Assembly Machine. Deep-learning-as …

Multi-Task Learning for Document Ranking and Query Suggestion

Title Multi-Task Learning for Document Ranking and Query Suggestion
Authors Wasi Uddin Ahmad, Kai-Wei Chang, Hongning Wang
Abstract We propose a multi-task learning framework to jointly learn document ranking and query suggestion for web search. It consists of two major components, a document ranker, and a query recommender. Document ranker combines current query and session information and compares the combined representation with document representation to rank the documents. Query recommender tracks users’ query reformulation sequence considering all previous in-session queries using a sequence to sequence approach. As both tasks are driven by the users’ underlying search intent, we perform joint learning of these two components through session recurrence, which encodes search context and intent. Extensive comparisons against state-of-the-art document ranking and query suggestion algorithms are performed on the public AOL search log, and the promising results endorse the effectiveness of the joint learning framework.
Tasks Document Ranking, Multi-Task Learning
Published 2018-01-01
URL https://openreview.net/forum?id=SJ1nzBeA-
PDF https://openreview.net/pdf?id=SJ1nzBeA-
PWC https://paperswithcode.com/paper/multi-task-learning-for-document-ranking-and
Repo https://github.com/wasiahmad/mnsrf_ranking_suggestion
Framework pytorch

Image Deblurring with a Class-Specific Prior

Title Image Deblurring with a Class-Specific Prior
Authors Saeed Anwar ; Cong Phuoc Huynh ; Fatih Porikli
Abstract A fundamental problem in image deblurring is to recover reliably distinct spatial frequencies that have been suppressed by the blur kernel. To tackle this issue, existing image deblurring techniques often rely on generic image priors such as the sparsity of salient features including image gradients and edges. However, these priors only help recover part of the frequency spectrum, such as the frequencies near the high-end. To this end, we pose the following specific questions: (i) Does any image class information offer an advantage over existing generic priors for image quality restoration? (ii) If a class-specific prior exists, how should it be encoded into a deblurring framework to recover attenuated image frequencies? Throughout this work, we devise a class-specific prior based on the band-pass filter responses and incorporate it into a deblurring strategy. More specifically, we show that the subspace of band-pass filtered images and their intensity distributions serve as useful priors for recovering image frequencies that are difficult to recover by generic image priors. We demonstrate that our image deblurring framework, when equipped with the above priors, significantly outperforms many state-of-the-art methods using generic image priors or class-specific exemplars.
Tasks Deblurring
Published 2018-07-11
URL https://ieeexplore.ieee.org/document/8409975
PDF https://ieeexplore.ieee.org/document/8409975
PWC https://paperswithcode.com/paper/image-deblurring-with-a-class-specific-prior
Repo https://github.com/saeed-anwar/Class_Specific_Deblurring
Framework none

Identifying Key Sentences for Precision Oncology Using Semi-Supervised Learning

Title Identifying Key Sentences for Precision Oncology Using Semi-Supervised Learning
Authors Jurica {\v{S}}eva, Martin Wackerbauer, Ulf Leser
Abstract We present a machine learning pipeline that identifies key sentences in abstracts of oncological articles to aid evidence-based medicine. This problem is characterized by the lack of gold standard datasets, data imbalance and thematic differences between available silver standard corpora. Additionally, available training and target data differs with regard to their domain (professional summaries vs. sentences in abstracts). This makes supervised machine learning inapplicable. We propose the use of two semi-supervised machine learning approaches: To mitigate difficulties arising from heterogeneous data sources, overcome data imbalance and create reliable training data we propose using transductive learning from positive and unlabelled data (PU Learning). For obtaining a realistic classification model, we propose the use of abstracts summarised in relevant sentences as unlabelled examples through Self-Training. The best model achieves 84{%} accuracy and 0.84 F1 score on our dataset
Tasks
Published 2018-07-01
URL https://www.aclweb.org/anthology/W18-2305/
PDF https://www.aclweb.org/anthology/W18-2305
PWC https://paperswithcode.com/paper/identifying-key-sentences-for-precision
Repo https://github.com/nachne/semisuper
Framework none

Introducing a Lexicon of Verbal Polarity Shifters for English

Title Introducing a Lexicon of Verbal Polarity Shifters for English
Authors Marc Schulder, Michael Wiegand, Josef Ruppenhofer, Stephanie Köser
Abstract
Tasks Natural Language Inference, Relation Extraction, Sentiment Analysis
Published 2018-05-01
URL https://www.aclweb.org/anthology/papers/L18-1222/l18-1222
PDF https://www.aclweb.org/anthology/L18-1222
PWC https://paperswithcode.com/paper/introducing-a-lexicon-of-verbal-polarity
Repo https://github.com/marcschulder/lrec2018
Framework none

Fast Adaptive RNN Encoder-Decoder for Anomaly Detection in SMD Assembly Machine

Title Fast Adaptive RNN Encoder-Decoder for Anomaly Detection in SMD Assembly Machine
Authors YeongHyeon Park, Il Dong Yun
Abstract Surface Mounted Device (SMD) assembly machine manufactures various products on a flexible manufacturing line. An anomaly detection model that can adapt to the various manufacturing environments very fast is required. In this paper, we proposed a fast adaptive anomaly detection model based on a Recurrent Neural Network (RNN) Encoder–Decoder with operating machine sounds. RNN Encoder–Decoder has a structure very similar to Auto-Encoder (AE), but the former has significantly reduced parameters compared to the latter because of its rolled structure. Thus, the RNN Encoder–Decoder only requires a short training process for fast adaptation. The anomaly detection model decides abnormality based on Euclidean distance between generated sequences and observed sequence from machine sounds. Experimental evaluation was conducted on a set of dataset from the SMD assembly machine. Results showed cutting-edge performance with fast adaptation. View Full-Text
Tasks Anomaly Detection
Published 2018-10-22
URL https://www.mdpi.com/1424-8220/18/10/3573
PDF https://www.mdpi.com/1424-8220/18/10/3573/pdf
PWC https://paperswithcode.com/paper/fast-adaptive-rnn-encoder-decoder-for-anomaly
Repo https://github.com/YeongHyeon/FARED_for_Anomaly_Detection
Framework tf

Deep-learning-assisted diagnosis for knee magnetic resonance imaging: Development and retrospective validation of MRNet

Title Deep-learning-assisted diagnosis for knee magnetic resonance imaging: Development and retrospective validation of MRNet
Authors Nicholas Bien, Pranav Rajpurkar, Robyn L. Ball, Jeremy Irvin, Allison Park, Erik Jones, Michael Bereket, Bhavik N. Patel, Kristen W. Yeom, Katie Shpanskaya, Safwan Halabi, Evan Zucker, Gary Fanton, Derek F. Amanatullah, Christopher F. Beaulieu, Geoffrey M. Riley, Russell J. Stewart, Francis G. Blankenberg, David B. Larson, Ricky H. Jones, Curtis P. Langlotz, Andrew Y. Ng, Matthew P. Lungren
Abstract Magnetic resonance imaging (MRI) of the knee is the preferred method for diagnosing knee injuries. However, interpretation of knee MRI is time-intensive and subject to diagnostic error and variability. An automated system for interpreting knee MRI could prioritize highrisk patients and assist clinicians in making diagnoses. Deep learning methods, in being able to automatically learn layers of features, are well suited for modeling the complex relationships between medical images and their interpretations. In this study we developed a deep learning model for detecting general abnormalities and specific diagnoses (anterior cruciate ligament [ACL] tears and meniscal tears) on knee MRI exams. We then measured the effect of providing the model’s predictions to clinical experts during interpretation.
Tasks
Published 2018-11-27
URL https://journals.plos.org/plosmedicine/article?id=10.1371/journal.pmed.1002699
PDF https://journals.plos.org/plosmedicine/article/file?id=10.1371/journal.pmed.1002699&type=printable
PWC https://paperswithcode.com/paper/deep-learning-assisted-diagnosis-for-knee
Repo https://github.com/ahmedbesbes/mrnet
Framework pytorch

Accelerated Spectral Ranking

Title Accelerated Spectral Ranking
Authors Arpit Agarwal, Prathamesh Patil, Shivani Agarwal
Abstract The problem of rank aggregation from pairwise and multiway comparisons has a wide range of implications, ranging from recommendation systems to sports rankings to social choice. Some of the most popular algorithms for this problem come from the class of spectral ranking algorithms; these include the rank centrality (RC) algorithm for pairwise comparisons, which returns consistent estimates under the Bradley-Terry-Luce (BTL) model for pairwise comparisons (Negahban et al., 2017), and its generalization, the Luce spectral ranking (LSR) algorithm, which returns consistent estimates under the more general multinomial logit (MNL) model for multiway comparisons (Maystre & Grossglauser, 2015). In this paper, we design a provably faster spectral ranking algorithm, which we call accelerated spectral ranking (ASR), that is also consistent under the MNL/BTL models. Our accelerated algorithm is achieved by designing a random walk that has a faster mixing time than the random walks associated with previous algorithms. In addition to a faster algorithm, our results yield improved sample complexity bounds for recovery of the MNL/BTL parameters: to the best of our knowledge, we give the first general sample complexity bounds for recovering the parameters of the MNL model from multiway comparisons under any (connected) comparison graph (and improve significantly over previous bounds for the BTL model for pairwise comparisons). We also give a message-passing interpretation of our algorithm, which suggests a decentralized distributed implementation. Our experiments on several real-world and synthetic datasets confirm that our new ASR algorithm is indeed orders of magnitude faster than existing algorithms.
Tasks Recommendation Systems
Published 2018-07-01
URL https://icml.cc/Conferences/2018/Schedule?showEvent=1999
PDF http://proceedings.mlr.press/v80/agarwal18b/agarwal18b.pdf
PWC https://paperswithcode.com/paper/accelerated-spectral-ranking
Repo https://github.com/agarpit/asr
Framework none

FireSim: FPGA-Accelerated Cycle-Exact Scale-Out System Simulation in the Public Cloud

Title FireSim: FPGA-Accelerated Cycle-Exact Scale-Out System Simulation in the Public Cloud
Authors Sagar Karandikar, Howard Mao, Donggyu Kim, David Biancolin, Alon Amid, Dayeol Lee, Nathan Pemberton, Emmanuel Amaro, Colin Schmidt, Aditya Chopra, Qijing Huang, Kyle Kovacs, Borivoje Nikolic, Randy Katz, Jonathan Bachrach, Krste Asanovic
Abstract We present FireSim, an open-source simulation platform that enables cycle-exact microarchitectural simulation of large scale-out clusters by combining FPGA-accelerated simulation of silicon-proven RTL designs with a scalable, distributed network simulation. Unlike prior FPGA-accelerated simulation tools, FireSim runs on Amazon EC2 F1, a public cloud FPGA platform, which greatly improves usability, provides elasticity, and lowers the cost of large-scale FPGAbased experiments. We describe the design and implementation of FireSim and show how it can provide sufficient performance to run modern applications at scale, to enable true hardware-software co-design. As an example, we demonstrate automatically generating and deploying a target cluster of 1,024 3.2 GHz quad-core server nodes, each with 16 GB of DRAM, interconnected by a 200 Gbit/s network with 2 microsecond latency, which simulates at a 3.4 MHz processor clock rate (less than 1,000x slowdown over real-time). In aggregate, this FireSim instantiation simulates 4,096 cores and 16 TB of memory, runs ˜14 billion instructions per second, and harnesses 12.8 million dollars worth of FPGAs—at a total cost of only ˜$100 per simulation hour to the user. We present several examples to show how FireSim can be used to explore various research directions in warehouse-scale machine design, including modeling networks with high-bandwidth and low-latency, integrating arbitrary RTL designs for a variety of commodity and specialized datacenter nodes, and modeling a variety of datacenter organizations, as well as reusing the scale-out FireSim infrastructure to enable fast, massively parallel cycle-exact single-node microarchitectural experimentation.
Tasks
Published 2018-06-02
URL https://dl.acm.org/citation.cfm?id=3276543
PDF https://sagark.org/assets/pubs/firesim-isca2018.pdf
PWC https://paperswithcode.com/paper/firesim-fpga-accelerated-cycle-exact-scale
Repo https://github.com/firesim/firesim
Framework none

Attention in Convolutional LSTM for Gesture Recognition

Title Attention in Convolutional LSTM for Gesture Recognition
Authors Liang Zhang, Guangming Zhu, Lin Mei, Peiyi Shen, Syed Afaq Ali Shah, Mohammed Bennamoun
Abstract Convolutional long short-term memory (LSTM) networks have been widely used for action/gesture recognition, and different attention mechanisms have also been embedded into the LSTM or the convolutional LSTM (ConvLSTM) networks. Based on the previous gesture recognition architectures which combine the three-dimensional convolution neural network (3DCNN) and ConvLSTM, this paper explores the effects of attention mechanism in ConvLSTM. Several variants of ConvLSTM are evaluated: (a) Removing the convolutional structures of the three gates in ConvLSTM, (b) Applying the attention mechanism on the input of ConvLSTM, (c) Reconstructing the input and (d) output gates respectively with the modified channel-wise attention mechanism. The evaluation results demonstrate that the spatial convolutions in the three gates scarcely contribute to the spatiotemporal feature fusion, and the attention mechanisms embedded into the input and output gates cannot improve the feature fusion. In other words, ConvLSTM mainly contributes to the temporal fusion along with the recurrent steps to learn the long-term spatiotemporal features, when taking as input the spatial or spatiotemporal features. On this basis, a new variant of LSTM is derived, in which the convolutional structures are only embedded into the input-to-state transition of LSTM. The code of the LSTM variants is publicly available.
Tasks Gesture Recognition
Published 2018-12-01
URL http://papers.nips.cc/paper/7465-attention-in-convolutional-lstm-for-gesture-recognition
PDF http://papers.nips.cc/paper/7465-attention-in-convolutional-lstm-for-gesture-recognition.pdf
PWC https://paperswithcode.com/paper/attention-in-convolutional-lstm-for-gesture
Repo https://github.com/GuangmingZhu/AttentionConvLSTM
Framework tf

Studying Muslim Stereotyping through Microportrait Extraction

Title Studying Muslim Stereotyping through Microportrait Extraction
Authors Antske Fokkens, Nel Ruigrok, Camiel Beukeboom, Gagestein Sarah, Wouter van Atteveldt
Abstract
Tasks
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1590/
PDF https://www.aclweb.org/anthology/L18-1590
PWC https://paperswithcode.com/paper/studying-muslim-stereotyping-through
Repo https://github.com/cltl/micro-portraits
Framework none

Training and Adapting Multilingual NMT for Less-resourced and Morphologically Rich Languages

Title Training and Adapting Multilingual NMT for Less-resourced and Morphologically Rich Languages
Authors Mat{=\i}ss Rikters, M{=a}rcis Pinnis, Rihards Kri{\v{s}}lauks
Abstract
Tasks Machine Translation
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1595/
PDF https://www.aclweb.org/anthology/L18-1595
PWC https://paperswithcode.com/paper/training-and-adapting-multilingual-nmt-for
Repo https://github.com/tilde-nlp/multilingual-nmt-data-prep
Framework none

mlpack 3: a fast, flexible machine learning library

Title mlpack 3: a fast, flexible machine learning library
Authors Ryan R. Curtin, Marcus Edel, Mikhail Lozhnikov, Yannis Mentekidis, Sumedh Ghaisas, Shangtong Zhang
Abstract In the past several years, the field of machine learning has seen an explosion of interest and excitement, with hundreds or thousands of algorithms developed for different tasks every year. But a primary problem faced by the field is the ability to scale to larger and larger data – since it is known that training on larger datasets typically produces better results. Therefore, the development of new algorithms for the continued growth of the field depends largely on the existence of good tooling and libraries that enable researchers and practitioners to quickly prototype and develop solutions. Simultaneously, useful libraries must also be efficient and well-implemented. This has motivated our development of mlpack. mlpack is a flexible and fast machine learning library written in C++ that has bindings that allow use from the command-line and from Python, with support for other languages in active development. mlpack has been developed actively for over 10 years, with over 100 contributors from around the world, and is a frequent mentoring organization in the Google Summer of Code program. If used in C++, the library allows flexibility with no speed penalty through policy-based design and template metaprogramming; but bindings are available to other languages, which allow easy use of the fast mlpack codebase. For fast linear algebra, mlpack is built on the Armadillo C++ matrix library, which in turn can use an optimized BLAS implementation such as OpenBLAS or even NVBLAS which would allow mlpack algorithms to be run on the GPU. In order to provide fast code, template metaprogramming is used throughout the library to reduce runtime overhead by performing any possible computations and optimizations at compile time. An automatic benchmarking system is developed and used to test the efficiency of mlpack’s algorithms. mlpack contains a number of standard machine learning algorithms, such as logistic regression, random forests, and k-means clustering, and also contains cutting-edge techniques such as a compile-time optimized deep learning and reinforcement learning framework, dual-tree algorithms for nearest neighbor search and other tasks, a generic optimization framework with numerous optimizers, a generic hyper-parameter tuner, and other recently published machine learning algorithms. For a more comprehensive introduction to mlpack, see the website at http://www.mlpack.org/
Tasks
Published 2018-06-18
URL https://doi.org/10.21105/joss.00726
PDF https://doi.org/10.21105/joss.00726
PWC https://paperswithcode.com/paper/mlpack-3-a-fast-flexible-machine-learning
Repo https://github.com/mlpack/mlpack
Framework none

Memory Replay GANs: Learning to Generate New Categories without Forgetting

Title Memory Replay GANs: Learning to Generate New Categories without Forgetting
Authors Chenshen Wu, Luis Herranz, Xialei Liu, Yaxing Wang, Joost Van De Weijer, Bogdan Raducanu
Abstract Previous works on sequential learning address the problem of forgetting in discriminative models. In this paper we consider the case of generative models. In particular, we investigate generative adversarial networks (GANs) in the task of learning new categories in a sequential fashion. We first show that sequential fine tuning renders the network unable to properly generate images from previous categories (i.e. forgetting). Addressing this problem, we propose Memory Replay GANs (MeRGANs), a conditional GAN framework that integrates a memory replay generator. We study two methods to prevent forgetting by leveraging these replays, namely joint training with replay and replay alignment. Qualitative and quantitative experimental results in MNIST, SVHN and LSUN datasets show that our memory replay approach can generate competitive images while significantly mitigating the forgetting of previous categories.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/7836-memory-replay-gans-learning-to-generate-new-categories-without-forgetting
PDF http://papers.nips.cc/paper/7836-memory-replay-gans-learning-to-generate-new-categories-without-forgetting.pdf
PWC https://paperswithcode.com/paper/memory-replay-gans-learning-to-generate-new
Repo https://github.com/WuChenshen/MeRGAN
Framework tf

A review of Spanish corpora annotated with negation

Title A review of Spanish corpora annotated with negation
Authors Salud Mar{'\i}a Jim{'e}nez-Zafra, Roser Morante, Maite Martin, L. Alfonso Ure{~n}a-L{'o}pez
Abstract The availability of corpora annotated with negation information is essential to develop negation processing systems in any language. However, there is a lack of these corpora even for languages like English, and when there are corpora available they are small and the annotations are not always compatible across corpora. In this paper we review the existing corpora annotated with negation in Spanish with the purpose of first, gathering the information to make it available for other researchers and, second, analyzing how compatible are the corpora and how has the linguistic phenomenon been addressed. Our final aim is to develop a supervised negation processing system for Spanish, for which we need training and test data. Our analysis shows that it will not be possible to merge the small corpora existing for Spanish due to lack of compatibility in the annotations.
Tasks Question Answering, Sentiment Analysis
Published 2018-08-01
URL https://www.aclweb.org/anthology/C18-1078/
PDF https://www.aclweb.org/anthology/C18-1078
PWC https://paperswithcode.com/paper/a-review-of-spanish-corpora-annotated-with
Repo https://github.com/sjzafra/spanish_negation_corpora
Framework none

Data Anonymization for Requirements Quality Analysis: a Reproducible Automatic Error Detection Task

Title Data Anonymization for Requirements Quality Analysis: a Reproducible Automatic Error Detection Task
Authors Juyeon Kang, Jungyeul Park
Abstract
Tasks
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1702/
PDF https://www.aclweb.org/anthology/L18-1702
PWC https://paperswithcode.com/paper/data-anonymization-for-requirements-quality
Repo https://github.com/jungyeul/rqa
Framework none
comments powered by Disqus