October 15, 2019

2748 words 13 mins read

Paper Group NANR 245

Paper Group NANR 245

Spot the Odd Man Out: Exploring the Associative Power of Lexical Resources. Towards Opinion Summarization of Customer Reviews. Voice Imitating Text-to-Speech Neural Networks. GraphBit: Bitwise Interaction Mining via Deep Reinforcement Learning. Neural Networks for irregularly observed continuous-time Stochastic Processes. Learning to Optimize Combi …

Spot the Odd Man Out: Exploring the Associative Power of Lexical Resources

Title Spot the Odd Man Out: Exploring the Associative Power of Lexical Resources
Authors Gabriel Stanovsky, Mark Hopkins
Abstract We propose Odd-Man-Out, a novel task which aims to test different properties of word representations. An Odd-Man-Out puzzle is composed of 5 (or more) words, and requires the system to choose the one which does not belong with the others. We show that this simple setup is capable of teasing out various properties of different popular lexical resources (like WordNet and pre-trained word embeddings), while being intuitive enough to annotate on a large scale. In addition, we propose a novel technique for training multi-prototype word representations, based on unsupervised clustering of ELMo embeddings, and show that it surpasses all other representations on all Odd-Man-Out collections.
Tasks Natural Language Inference, Question Answering, Word Embeddings
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1182/
PDF https://www.aclweb.org/anthology/D18-1182
PWC https://paperswithcode.com/paper/spot-the-odd-man-out-exploring-the
Repo
Framework

Towards Opinion Summarization of Customer Reviews

Title Towards Opinion Summarization of Customer Reviews
Authors Samuel Pecar
Abstract In recent years, the number of texts has grown rapidly. For example, most review-based portals, like Yelp or Amazon, contain thousands of user-generated reviews. It is impossible for any human reader to process even the most relevant of these documents. The most promising tool to solve this task is a text summarization. Most existing approaches, however, work on small, homogeneous, English datasets, and do not account to multi-linguality, opinion shift, and domain effects. In this paper, we introduce our research plan to use neural networks on user-generated travel reviews to generate summaries that take into account shifting opinions over time. We outline future directions in summarization to address all of these issues. By resolving the existing problems, we will make it easier for users of review-sites to make more informed decisions.
Tasks Abstractive Text Summarization, Decision Making, Text Summarization
Published 2018-07-01
URL https://www.aclweb.org/anthology/P18-3001/
PDF https://www.aclweb.org/anthology/P18-3001
PWC https://paperswithcode.com/paper/towards-opinion-summarization-of-customer
Repo
Framework

Voice Imitating Text-to-Speech Neural Networks

Title Voice Imitating Text-to-Speech Neural Networks
Authors Younggun Lee, Taesu Kim, Soo-Young Lee
Abstract We propose a neural text-to-speech (TTS) model that can imitate a new speaker’s voice using only a small amount of speech sample. We demonstrate voice imitation using only a 6-seconds long speech sample without any other information such as transcripts. Our model also enables voice imitation instantly without additional training of the model. We implemented the voice imitating TTS model by combining a speaker embedder network with a state-of-the-art TTS model, Tacotron. The speaker embedder network takes a new speaker’s speech sample and returns a speaker embedding. The speaker embedding with a target sentence are fed to Tacotron, and speech is generated with the new speaker’s voice. We show that the speaker embeddings extracted by the speaker embedder network can represent the latent structure in different voices. The generated speech samples from our model have comparable voice quality to the ones from existing multi-speaker TTS models.
Tasks
Published 2018-06-04
URL https://arxiv.org/abs/1806.00927
PDF https://arxiv.org/pdf/1806.00927
PWC https://paperswithcode.com/paper/voice-imitating-text-to-speech-neural
Repo
Framework

GraphBit: Bitwise Interaction Mining via Deep Reinforcement Learning

Title GraphBit: Bitwise Interaction Mining via Deep Reinforcement Learning
Authors Yueqi Duan, Ziwei Wang, Jiwen Lu, Xudong Lin, Jie Zhou
Abstract In this paper, we propose a GraphBit method to learn deep binary descriptors in a directed acyclic graph unsupervisedly, representing bitwise interactions as edges between the nodes of bits. Conventional binary representation learning methods enforce each element to be binarized into zero or one. However, there are elements lying in the boundary which suffer from doubtful binarization as ``ambiguous bits’'. Ambiguous bits fail to collect effective information for confident binarization, which are unreliable and sensitive to noise. We argue that there are implicit inner relationships between bits in binary descriptors, where the related bits can provide extra instruction as prior knowledge for ambiguity elimination. Specifically, we design a deep reinforcement learning model to learn the structure of the graph for bitwise interaction mining, reducing the uncertainty of binary codes by maximizing the mutual information with inputs and related bits, so that the ambiguous bits receive additional instruction from the graph for confident binarization. Due to the reliability of the proposed binary codes with bitwise interaction, we obtain an average improvement of 9.64%, 8.84% and 3.22% on the CIFAR-10, Brown and HPatches datasets respectively compared with the state-of-the-art unsupervised binary descriptors. |
Tasks Representation Learning
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/Duan_GraphBit_Bitwise_Interaction_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/Duan_GraphBit_Bitwise_Interaction_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/graphbit-bitwise-interaction-mining-via-deep
Repo
Framework

Neural Networks for irregularly observed continuous-time Stochastic Processes

Title Neural Networks for irregularly observed continuous-time Stochastic Processes
Authors Francois W. Belletti, Alexander Ku, Joseph E. Gonzalez
Abstract Designing neural networks for continuous-time stochastic processes is challenging, especially when observations are made irregularly. In this article, we analyze neural networks from a frame theoretic perspective to identify the sufficient conditions that enable smoothly recoverable representations of signals in L^2(R). Moreover, we show that, under certain assumptions, these properties hold even when signals are irregularly observed. As we converge to the family of (convolutional) neural networks that satisfy these conditions, we show that we can optimize our convolution filters while constraining them so that they effectively compute a Discrete Wavelet Transform. Such a neural network can efficiently divide the time-axis of a signal into orthogonal sub-spaces of different temporal scale and localization. We evaluate the resulting neural network on an assortment of synthetic and real-world tasks: parsimonious auto-encoding, video classification, and financial forecasting.
Tasks Video Classification
Published 2018-01-01
URL https://openreview.net/forum?id=S1fHmlbCW
PDF https://openreview.net/pdf?id=S1fHmlbCW
PWC https://paperswithcode.com/paper/neural-networks-for-irregularly-observed
Repo
Framework

Learning to Optimize Combinatorial Functions

Title Learning to Optimize Combinatorial Functions
Authors Nir Rosenfeld, Eric Balkanski, Amir Globerson, Yaron Singer
Abstract Submodular functions have become a ubiquitous tool in machine learning. They are learnable from data, and can be optimized efficiently and with guarantees. Nonetheless, recent negative results show that optimizing learned surrogates of submodular functions can result in arbitrarily bad approximations of the true optimum. Our goal in this paper is to highlight the source of this hardness, and propose an alternative criterion for optimizing general combinatorial functions from sampled data. We prove a tight equivalence showing that a class of functions is optimizable if and only if it can be learned. We provide efficient and scalable optimization algorithms for several function classes of interest, and demonstrate their utility on the task of optimally choosing trending social media items.
Tasks
Published 2018-07-01
URL https://icml.cc/Conferences/2018/Schedule?showEvent=1881
PDF http://proceedings.mlr.press/v80/rosenfeld18a/rosenfeld18a.pdf
PWC https://paperswithcode.com/paper/learning-to-optimize-combinatorial-functions
Repo
Framework

Alpha-divergence bridges maximum likelihood and reinforcement learning in neural sequence generation

Title Alpha-divergence bridges maximum likelihood and reinforcement learning in neural sequence generation
Authors Sotetsu Koyamada, Yuta Kikuchi, Atsunori Kanemura, Shin-ichi Maeda, Shin Ishii
Abstract Neural sequence generation is commonly approached by using maximum- likelihood (ML) estimation or reinforcement learning (RL). However, it is known that they have their own shortcomings; ML presents training/testing discrepancy, whereas RL suffers from sample inefficiency. We point out that it is difficult to resolve all of the shortcomings simultaneously because of a tradeoff between ML and RL. In order to counteract these problems, we propose an objective function for sequence generation using α-divergence, which leads to an ML-RL integrated method that exploits better parts of ML and RL. We demonstrate that the proposed objective function generalizes ML and RL objective functions because it includes both as its special cases (ML corresponds to α → 0 and RL to α → 1). We provide a proposition stating that the difference between the RL objective function and the proposed one monotonically decreases with increasing α. Experimental results on machine translation tasks show that minimizing the proposed objective function achieves better sequence generation performance than ML-based methods.
Tasks Machine Translation
Published 2018-01-01
URL https://openreview.net/forum?id=H1Nyf7W0Z
PDF https://openreview.net/pdf?id=H1Nyf7W0Z
PWC https://paperswithcode.com/paper/alpha-divergence-bridges-maximum-likelihood
Repo
Framework

Keyphrases Extraction from User-Generated Contents in Healthcare Domain Using Long Short-Term Memory Networks

Title Keyphrases Extraction from User-Generated Contents in Healthcare Domain Using Long Short-Term Memory Networks
Authors Ilham Fathy Saputra, Rahmad Mahendra, Alfan Farizki Wicaksono
Abstract We propose keyphrases extraction technique to extract important terms from the healthcare user-generated contents. We employ deep learning architecture, i.e. Long Short-Term Memory, and leverage word embeddings, medical concepts from a knowledge base, and linguistic components as our features. The proposed model achieves 61.37{%} F-1 score. Experimental results indicate that our proposed approach outperforms the baseline methods, i.e. RAKE and CRF, on the task of extracting keyphrases from Indonesian health forum posts.
Tasks Question Answering, Text Classification, Text Summarization, Word Embeddings
Published 2018-07-01
URL https://www.aclweb.org/anthology/W18-2304/
PDF https://www.aclweb.org/anthology/W18-2304
PWC https://paperswithcode.com/paper/keyphrases-extraction-from-user-generated
Repo
Framework

Teaching Machines to Understand Baseball Games: Large-Scale Baseball Video Database for Multiple Video Understanding Tasks

Title Teaching Machines to Understand Baseball Games: Large-Scale Baseball Video Database for Multiple Video Understanding Tasks
Authors Minho Shim, Young Hwi Kim, Kyungmin Kim, Seon Joo Kim
Abstract A major obstacle in teaching machines to understand videos is the lack of training data, as creating temporal annotations for long videos requires a huge amount of human effort. To this end, we introduce a new large-scale baseball video dataset called the BBDB, which is produced semi-automatically by using play-by-play texts available online. The BBDB contains 4200 hours of baseball game videos with 400k temporally annotated activity segments. The new dataset has several major challenging factors compared to other datasets: 1) the dataset contains a large number of visually similar segments with different labels. 2) It can be used for many video understanding tasks including video recognition, localization, text-video alignment, video highlight generation, and data imbalance problem. To observe the potential of the BBDB, we conducted extensive experiments by running many different types of video understanding algorithms on our new dataset. The database is available at https://sites.google.com/site/eccv2018bbdb/
Tasks Video Alignment, Video Recognition, Video Understanding
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/Minho_Shim_Teaching_Machines_to_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/Minho_Shim_Teaching_Machines_to_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/teaching-machines-to-understand-baseball
Repo
Framework

Proportional Allocation: Simple, Distributed, and Diverse Matching with High Entropy

Title Proportional Allocation: Simple, Distributed, and Diverse Matching with High Entropy
Authors Shipra Agrawal, Morteza Zadimoghaddam, Vahab Mirrokni
Abstract Inspired by many applications of bipartite matching in online advertising and machine learning, we study a simple and natural iterative proportional allocation algorithm: Maintain a priority score $\priority_a$ for each node $a\in \mathds{A}$ on one side of the bipartition, initialized as $\priority_a=1$. Iteratively allocate the nodes $i\in \impressions$ on the other side to eligible nodes in $\mathds{A}$ in proportion of their priority scores. After each round, for each node $a\in \mathds{A}$, decrease or increase the score $\priority_a$ based on whether it is over- or under- allocated. Our first result is that this simple, distributed algorithm converges to a $(1-\epsilon)$-approximate fractional $b$-matching solution in $O({\log n\over \epsilon^2} )$ rounds. We also extend the proportional allocation algorithm and convergence results to the maximum weighted matching problem, and show that the algorithm can be naturally tuned to produce maximum matching with high entropy. High entropy, in turn, implies additional desirable properties of this matching, e.g., it satisfies certain diversity and fairness (aka anonymity) properties that are desirable in a variety of applications in online advertising and machine learning.
Tasks
Published 2018-07-01
URL https://icml.cc/Conferences/2018/Schedule?showEvent=2388
PDF http://proceedings.mlr.press/v80/agrawal18b/agrawal18b.pdf
PWC https://paperswithcode.com/paper/proportional-allocation-simple-distributed
Repo
Framework

Fast and Accurate Text Classification: Skimming, Rereading and Early Stopping

Title Fast and Accurate Text Classification: Skimming, Rereading and Early Stopping
Authors Keyi Yu, Yang Liu, Alexander G. Schwing, Jian Peng
Abstract Recent advances in recurrent neural nets (RNNs) have shown much promise in many applications in natural language processing. For most of these tasks, such as sentiment analysis of customer reviews, a recurrent neural net model parses the entire review before forming a decision. We argue that reading the entire input is not always necessary in practice, since a lot of reviews are often easy to classify, i.e., a decision can be formed after reading some crucial sentences or words in the provided text. In this paper, we present an approach of fast reading for text classification. Inspired by several well-known human reading techniques, our approach implements an intelligent recurrent agent which evaluates the importance of the current snippet in order to decide whether to make a prediction, or to skip some texts, or to re-read part of the sentence. Our agent uses an RNN module to encode information from the past and the current tokens, and applies a policy module to form decisions. With an end-to-end training algorithm based on policy gradient, we train and test our agent on several text classification datasets and achieve both higher efficiency and better accuracy compared to previous approaches.
Tasks Sentiment Analysis, Text Classification
Published 2018-01-01
URL https://openreview.net/forum?id=ryZ8sz-Ab
PDF https://openreview.net/pdf?id=ryZ8sz-Ab
PWC https://paperswithcode.com/paper/fast-and-accurate-text-classification
Repo
Framework

Practical exact algorithm for trembling-hand equilibrium refinements in games

Title Practical exact algorithm for trembling-hand equilibrium refinements in games
Authors Gabriele Farina, Nicola Gatti, Tuomas Sandholm
Abstract Nash equilibrium strategies have the known weakness that they do not prescribe rational play in situations that are reached with zero probability according to the strategies themselves, for example, if players have made mistakes. Trembling-hand refinements—such as extensive-form perfect equilibria and quasi-perfect equilibria—remedy this problem in sound ways. Despite their appeal, they have not received attention in practice since no known algorithm for computing them scales beyond toy instances. In this paper, we design an exact polynomial-time algorithm for finding trembling-hand equilibria in zero-sum extensive-form games. It is several orders of magnitude faster than the best prior ones, numerically stable, and quickly solves game instances with tens of thousands of nodes in the game tree. It enables, for the first time, the use of trembling-hand refinements in practice.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/7751-practical-exact-algorithm-for-trembling-hand-equilibrium-refinements-in-games
PDF http://papers.nips.cc/paper/7751-practical-exact-algorithm-for-trembling-hand-equilibrium-refinements-in-games.pdf
PWC https://paperswithcode.com/paper/practical-exact-algorithm-for-trembling-hand
Repo
Framework

Controlling Length in Abstractive Summarization Using a Convolutional Neural Network

Title Controlling Length in Abstractive Summarization Using a Convolutional Neural Network
Authors Yizhu Liu, Zhiyi Luo, Kenny Zhu
Abstract Convolutional neural networks (CNNs) have met great success in abstractive summarization, but they cannot effectively generate summaries of desired lengths. Because generated summaries are used in difference scenarios which may have space or length constraints, the ability to control the summary length in abstractive summarization is an important problem. In this paper, we propose an approach to constrain the summary length by extending a convolutional sequence to sequence model. The results show that this approach generates high-quality summaries with user defined length, and outperforms the baselines consistently in terms of ROUGE score, length variations and semantic similarity.
Tasks Abstractive Text Summarization, Machine Translation, Semantic Similarity, Semantic Textual Similarity
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1444/
PDF https://www.aclweb.org/anthology/D18-1444
PWC https://paperswithcode.com/paper/controlling-length-in-abstractive
Repo
Framework
Title Learning Representations for Faster Similarity Search
Authors Ludwig Schmidt, Kunal Talwar
Abstract In high dimensions, the performance of nearest neighbor algorithms depends crucially on structure in the data. While traditional nearest neighbor datasets consisted mostly of hand-crafted feature vectors, an increasing number of datasets comes from representations learned with neural networks. We study the interaction between nearest neighbor algorithms and neural networks in more detail. We find that the network architecture can significantly influence the efficacy of nearest neighbor algorithms even when the classification accuracy is unchanged. Based on our experiments, we propose a number of training modifications that lead to significantly better datasets for nearest neighbor algorithms. Our modifications lead to learned representations that can accelerate nearest neighbor queries by 5x.
Tasks
Published 2018-01-01
URL https://openreview.net/forum?id=SkrHeXbCW
PDF https://openreview.net/pdf?id=SkrHeXbCW
PWC https://paperswithcode.com/paper/learning-representations-for-faster
Repo
Framework

Fast Greedy MAP Inference for Determinantal Point Process to Improve Recommendation Diversity

Title Fast Greedy MAP Inference for Determinantal Point Process to Improve Recommendation Diversity
Authors Laming Chen, Guoxin Zhang, Eric Zhou
Abstract The determinantal point process (DPP) is an elegant probabilistic model of repulsion with applications in various machine learning tasks including summarization and search. However, the maximum a posteriori (MAP) inference for DPP which plays an important role in many applications is NP-hard, and even the popular greedy algorithm can still be too computationally expensive to be used in large-scale real-time scenarios. To overcome the computational challenge, in this paper, we propose a novel algorithm to greatly accelerate the greedy MAP inference for DPP. In addition, our algorithm also adapts to scenarios where the repulsion is only required among nearby few items in the result sequence. We apply the proposed algorithm to generate relevant and diverse recommendations. Experimental results show that our proposed algorithm is significantly faster than state-of-the-art competitors, and provides a better relevance-diversity trade-off on several public datasets, which is also confirmed in an online A/B test.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/7805-fast-greedy-map-inference-for-determinantal-point-process-to-improve-recommendation-diversity
PDF http://papers.nips.cc/paper/7805-fast-greedy-map-inference-for-determinantal-point-process-to-improve-recommendation-diversity.pdf
PWC https://paperswithcode.com/paper/fast-greedy-map-inference-for-determinantal
Repo
Framework
comments powered by Disqus