Paper Group AWR 140
Toward Diverse Text Generation with Inverse Reinforcement Learning. Multi-Agent Reinforcement Learning: A Report on Challenges and Approaches. Hierarchical Graph Clustering using Node Pair Sampling. Mem2Seq: Effectively Incorporating Knowledge Bases into End-to-End Task-Oriented Dialog Systems. Norm-Preservation: Why Residual Networks Can Become Ex …
Toward Diverse Text Generation with Inverse Reinforcement Learning
Title | Toward Diverse Text Generation with Inverse Reinforcement Learning |
Authors | Zhan Shi, Xinchi Chen, Xipeng Qiu, Xuanjing Huang |
Abstract | Text generation is a crucial task in NLP. Recently, several adversarial generative models have been proposed to improve the exposure bias problem in text generation. Though these models gain great success, they still suffer from the problems of reward sparsity and mode collapse. In order to address these two problems, in this paper, we employ inverse reinforcement learning (IRL) for text generation. Specifically, the IRL framework learns a reward function on training data, and then an optimal policy to maximum the expected total reward. Similar to the adversarial models, the reward and policy function in IRL are optimized alternately. Our method has two advantages: (1) the reward function can produce more dense reward signals. (2) the generation policy, trained by “entropy regularized” policy gradient, encourages to generate more diversified texts. Experiment results demonstrate that our proposed method can generate higher quality texts than the previous methods. |
Tasks | Text Generation |
Published | 2018-04-30 |
URL | http://arxiv.org/abs/1804.11258v3 |
http://arxiv.org/pdf/1804.11258v3.pdf | |
PWC | https://paperswithcode.com/paper/toward-diverse-text-generation-with-inverse |
Repo | https://github.com/FudanNLP/Irl_gen |
Framework | tf |
Multi-Agent Reinforcement Learning: A Report on Challenges and Approaches
Title | Multi-Agent Reinforcement Learning: A Report on Challenges and Approaches |
Authors | Sanyam Kapoor |
Abstract | Reinforcement Learning (RL) is a learning paradigm concerned with learning to control a system so as to maximize an objective over the long term. This approach to learning has received immense interest in recent times and success manifests itself in the form of human-level performance on games like \textit{Go}. While RL is emerging as a practical component in real-life systems, most successes have been in Single Agent domains. This report will instead specifically focus on challenges that are unique to Multi-Agent Systems interacting in mixed cooperative and competitive environments. The report concludes with advances in the paradigm of training Multi-Agent Systems called \textit{Decentralized Actor, Centralized Critic}, based on an extension of MDPs called \textit{Decentralized Partially Observable MDP}s, which has seen a renewed interest lately. |
Tasks | Multi-agent Reinforcement Learning |
Published | 2018-07-25 |
URL | http://arxiv.org/abs/1807.09427v1 |
http://arxiv.org/pdf/1807.09427v1.pdf | |
PWC | https://paperswithcode.com/paper/multi-agent-reinforcement-learning-a-report |
Repo | https://github.com/activatedgeek/marl-challenges-approaches |
Framework | none |
Hierarchical Graph Clustering using Node Pair Sampling
Title | Hierarchical Graph Clustering using Node Pair Sampling |
Authors | Thomas Bonald, Bertrand Charpentier, Alexis Galland, Alexandre Hollocou |
Abstract | We present a novel hierarchical graph clustering algorithm inspired by modularity-based clustering techniques. The algorithm is agglomerative and based on a simple distance between clusters induced by the probability of sampling node pairs. We prove that this distance is reducible, which enables the use of the nearest-neighbor chain to speed up the agglomeration. The output of the algorithm is a regular dendrogram, which reveals the multi-scale structure of the graph. The results are illustrated on both synthetic and real datasets. |
Tasks | Graph Clustering |
Published | 2018-06-05 |
URL | http://arxiv.org/abs/1806.01664v2 |
http://arxiv.org/pdf/1806.01664v2.pdf | |
PWC | https://paperswithcode.com/paper/hierarchical-graph-clustering-using-node-pair |
Repo | https://github.com/tbonald/paris |
Framework | none |
Mem2Seq: Effectively Incorporating Knowledge Bases into End-to-End Task-Oriented Dialog Systems
Title | Mem2Seq: Effectively Incorporating Knowledge Bases into End-to-End Task-Oriented Dialog Systems |
Authors | Andrea Madotto, Chien-Sheng Wu, Pascale Fung |
Abstract | End-to-end task-oriented dialog systems usually suffer from the challenge of incorporating knowledge bases. In this paper, we propose a novel yet simple end-to-end differentiable model called memory-to-sequence (Mem2Seq) to address this issue. Mem2Seq is the first neural generative model that combines the multi-hop attention over memories with the idea of pointer network. We empirically show how Mem2Seq controls each generation step, and how its multi-hop attention mechanism helps in learning correlations between memories. In addition, our model is quite general without complicated task-specific designs. As a result, we show that Mem2Seq can be trained faster and attain the state-of-the-art performance on three different task-oriented dialog datasets. |
Tasks | |
Published | 2018-04-23 |
URL | http://arxiv.org/abs/1804.08217v3 |
http://arxiv.org/pdf/1804.08217v3.pdf | |
PWC | https://paperswithcode.com/paper/mem2seq-effectively-incorporating-knowledge |
Repo | https://github.com/HLTCHKUST/Mem2Seq |
Framework | pytorch |
Norm-Preservation: Why Residual Networks Can Become Extremely Deep?
Title | Norm-Preservation: Why Residual Networks Can Become Extremely Deep? |
Authors | Alireza Zaeemzadeh, Nazanin Rahnavard, Mubarak Shah |
Abstract | Augmenting neural networks with skip connections, as introduced in the so-called ResNet architecture, surprised the community by enabling the training of networks of more than 1,000 layers with significant performance gains. This paper deciphers ResNet by analyzing the effect of skip connections, and puts forward new theoretical results on the advantages of identity skip connections in neural networks. We prove that the skip connections in the residual blocks facilitate preserving the norm of the gradient, and lead to stable back-propagation, which is desirable from optimization perspective. We also show that, perhaps surprisingly, as more residual blocks are stacked, the norm-preservation of the network is enhanced. Our theoretical arguments are supported by extensive empirical evidence. Can we push for extra norm-preservation? We answer this question by proposing an efficient method to regularize the singular values of the convolution operator and making the ResNet’s transition layers extra norm-preserving. Our numerical investigations demonstrate that the learning dynamics and the classification performance of ResNet can be improved by making it even more norm preserving. Our results and the introduced modification for ResNet, referred to as Procrustes ResNets, can be used as a guide for training deeper networks and can also inspire new deeper architectures. |
Tasks | |
Published | 2018-05-18 |
URL | https://arxiv.org/abs/1805.07477v4 |
https://arxiv.org/pdf/1805.07477v4.pdf | |
PWC | https://paperswithcode.com/paper/norm-preservation-why-residual-networks-can |
Repo | https://github.com/zaeemzadeh/ProcResNet |
Framework | pytorch |
The WiLI benchmark dataset for written language identification
Title | The WiLI benchmark dataset for written language identification |
Authors | Martin Thoma |
Abstract | This paper describes the WiLI-2018 benchmark dataset for monolingual written natural language identification. WiLI-2018 is a publicly available, free of charge dataset of short text extracts from Wikipedia. It contains 1000 paragraphs of 235 languages, totaling in 23500 paragraphs. WiLI is a classification dataset: Given an unknown paragraph written in one dominant language, it has to be decided which language it is. |
Tasks | Language Identification |
Published | 2018-01-23 |
URL | http://arxiv.org/abs/1801.07779v1 |
http://arxiv.org/pdf/1801.07779v1.pdf | |
PWC | https://paperswithcode.com/paper/the-wili-benchmark-dataset-for-written |
Repo | https://github.com/LauraRuis/LanguageIdentification |
Framework | pytorch |
Heated-Up Softmax Embedding
Title | Heated-Up Softmax Embedding |
Authors | Xu Zhang, Felix Xinnan Yu, Svebor Karaman, Wei Zhang, Shih-Fu Chang |
Abstract | Metric learning aims at learning a distance which is consistent with the semantic meaning of the samples. The problem is generally solved by learning an embedding for each sample such that the embeddings of samples of the same category are compact while the embeddings of samples of different categories are spread-out in the feature space. We study the features extracted from the second last layer of a deep neural network based classifier trained with the cross entropy loss on top of the softmax layer. We show that training classifiers with different temperature values of softmax function leads to features with different levels of compactness. Leveraging these insights, we propose a “heating-up” strategy to train a classifier with increasing temperatures, leading the corresponding embeddings to achieve state-of-the-art performance on a variety of metric learning benchmarks. |
Tasks | Metric Learning |
Published | 2018-09-11 |
URL | http://arxiv.org/abs/1809.04157v1 |
http://arxiv.org/pdf/1809.04157v1.pdf | |
PWC | https://paperswithcode.com/paper/heated-up-softmax-embedding |
Repo | https://github.com/ColumbiaDVMM/Heated_Up_Softmax_Embedding |
Framework | tf |
Graph2Seq: Graph to Sequence Learning with Attention-based Neural Networks
Title | Graph2Seq: Graph to Sequence Learning with Attention-based Neural Networks |
Authors | Kun Xu, Lingfei Wu, Zhiguo Wang, Yansong Feng, Michael Witbrock, Vadim Sheinin |
Abstract | The celebrated Sequence to Sequence learning (Seq2Seq) technique and its numerous variants achieve excellent performance on many tasks. However, many machine learning tasks have inputs naturally represented as graphs; existing Seq2Seq models face a significant challenge in achieving accurate conversion from graph form to the appropriate sequence. To address this challenge, we introduce a novel general end-to-end graph-to-sequence neural encoder-decoder model that maps an input graph to a sequence of vectors and uses an attention-based LSTM method to decode the target sequence from these vectors. Our method first generates the node and graph embeddings using an improved graph-based neural network with a novel aggregation strategy to incorporate edge direction information in the node embeddings. We further introduce an attention mechanism that aligns node embeddings and the decoding sequence to better cope with large graphs. Experimental results on bAbI, Shortest Path, and Natural Language Generation tasks demonstrate that our model achieves state-of-the-art performance and significantly outperforms existing graph neural networks, Seq2Seq, and Tree2Seq models; using the proposed bi-directional node embedding aggregation strategy, the model can converge rapidly to the optimal performance. |
Tasks | Graph-to-Sequence, SQL-to-Text, Text Generation |
Published | 2018-04-03 |
URL | http://arxiv.org/abs/1804.00823v4 |
http://arxiv.org/pdf/1804.00823v4.pdf | |
PWC | https://paperswithcode.com/paper/graph2seq-graph-to-sequence-learning-with |
Repo | https://github.com/IBM/Graph2Seq |
Framework | tf |
End-to-End Video Captioning with Multitask Reinforcement Learning
Title | End-to-End Video Captioning with Multitask Reinforcement Learning |
Authors | Lijun Li, Boqing Gong |
Abstract | Although end-to-end (E2E) learning has led to impressive progress on a variety of visual understanding tasks, it is often impeded by hardware constraints (e.g., GPU memory) and is prone to overfitting. When it comes to video captioning, one of the most challenging benchmark tasks in computer vision, those limitations of E2E learning are especially amplified by the fact that both the input videos and output captions are lengthy sequences. Indeed, state-of-the-art methods for video captioning process video frames by convolutional neural networks and generate captions by unrolling recurrent neural networks. If we connect them in an E2E manner, the resulting model is both memory-consuming and data-hungry, making it extremely hard to train. In this paper, we propose a multitask reinforcement learning approach to training an E2E video captioning model. The main idea is to mine and construct as many effective tasks (e.g., attributes, rewards, and the captions) as possible from the human captioned videos such that they can jointly regulate the search space of the E2E neural network, from which an E2E video captioning model can be found and generalized to the testing phase. To the best of our knowledge, this is the first video captioning model that is trained end-to-end from the raw video input to the caption output. Experimental results show that such a model outperforms existing ones to a large margin on two benchmark video captioning datasets. |
Tasks | Video Captioning |
Published | 2018-03-21 |
URL | http://arxiv.org/abs/1803.07950v2 |
http://arxiv.org/pdf/1803.07950v2.pdf | |
PWC | https://paperswithcode.com/paper/end-to-end-video-captioning-with-multitask |
Repo | https://github.com/adwardlee/multitask-end-to-end-video-captioning |
Framework | tf |
Near-Optimal Representation Learning for Hierarchical Reinforcement Learning
Title | Near-Optimal Representation Learning for Hierarchical Reinforcement Learning |
Authors | Ofir Nachum, Shixiang Gu, Honglak Lee, Sergey Levine |
Abstract | We study the problem of representation learning in goal-conditioned hierarchical reinforcement learning. In such hierarchical structures, a higher-level controller solves tasks by iteratively communicating goals which a lower-level policy is trained to reach. Accordingly, the choice of representation – the mapping of observation space to goal space – is crucial. To study this problem, we develop a notion of sub-optimality of a representation, defined in terms of expected reward of the optimal hierarchical policy using this representation. We derive expressions which bound the sub-optimality and show how these expressions can be translated to representation learning objectives which may be optimized in practice. Results on a number of difficult continuous-control tasks show that our approach to representation learning yields qualitatively better representations as well as quantitatively better hierarchical policies, compared to existing methods (see videos at https://sites.google.com/view/representation-hrl). |
Tasks | Continuous Control, Hierarchical Reinforcement Learning, Representation Learning |
Published | 2018-10-02 |
URL | http://arxiv.org/abs/1810.01257v2 |
http://arxiv.org/pdf/1810.01257v2.pdf | |
PWC | https://paperswithcode.com/paper/near-optimal-representation-learning-for |
Repo | https://github.com/josherich/efficient-hrl |
Framework | tf |
Testing Deep Neural Networks
Title | Testing Deep Neural Networks |
Authors | Youcheng Sun, Xiaowei Huang, Daniel Kroening, James Sharp, Matthew Hill, Rob Ashmore |
Abstract | Deep neural networks (DNNs) have a wide range of applications, and software employing them must be thoroughly tested, especially in safety-critical domains. However, traditional software test coverage metrics cannot be applied directly to DNNs. In this paper, inspired by the MC/DC coverage criterion, we propose a family of four novel test criteria that are tailored to structural features of DNNs and their semantics. We validate the criteria by demonstrating that the generated test inputs guided via our proposed coverage criteria are able to capture undesired behaviours in a DNN. Test cases are generated using a symbolic approach and a gradient-based heuristic search. By comparing them with existing methods, we show that our criteria achieve a balance between their ability to find bugs (proxied using adversarial examples) and the computational cost of test case generation. Our experiments are conducted on state-of-the-art DNNs obtained using popular open source datasets, including MNIST, CIFAR-10 and ImageNet. |
Tasks | |
Published | 2018-03-10 |
URL | http://arxiv.org/abs/1803.04792v4 |
http://arxiv.org/pdf/1803.04792v4.pdf | |
PWC | https://paperswithcode.com/paper/testing-deep-neural-networks |
Repo | https://github.com/theyoucheng/deepcover |
Framework | none |
Fast Symbolic 3D Registration Solution
Title | Fast Symbolic 3D Registration Solution |
Authors | Jin Wu, Ming Liu, Zebo Zhou, Rui Li |
Abstract | 3D registration has always been performed invoking singular value decomposition (SVD) or eigenvalue decomposition (EIG) in real engineering practices. However, these numerical algorithms suffer from uncertainty of convergence in many cases. A novel fast symbolic solution is proposed in this paper by following our recent publication in this journal. The equivalence analysis shows that our previous solver can be converted to deal with the 3D registration problem. Rather, the computation procedure is studied for further simplification of computing without complex-number support. Experimental results show that the proposed solver does not loose accuracy and robustness but improves the execution speed to a large extent by almost %50 to %80, on both personal computer and embedded processor. |
Tasks | |
Published | 2018-05-12 |
URL | https://arxiv.org/abs/1805.08703v3 |
https://arxiv.org/pdf/1805.08703v3.pdf | |
PWC | https://paperswithcode.com/paper/fast-symbolic-3d-registration-solution |
Repo | https://github.com/zarathustr/FS3R-Matlab |
Framework | none |
Automatic Estimation of Simultaneous Interpreter Performance
Title | Automatic Estimation of Simultaneous Interpreter Performance |
Authors | Craig Stewart, Nikolai Vogler, Junjie Hu, Jordan Boyd-Graber, Graham Neubig |
Abstract | Simultaneous interpretation, translation of the spoken word in real-time, is both highly challenging and physically demanding. Methods to predict interpreter confidence and the adequacy of the interpreted message have a number of potential applications, such as in computer-assisted interpretation interfaces or pedagogical tools. We propose the task of predicting simultaneous interpreter performance by building on existing methodology for quality estimation (QE) of machine translation output. In experiments over five settings in three language pairs, we extend a QE pipeline to estimate interpreter performance (as approximated by the METEOR evaluation metric) and propose novel features reflecting interpretation strategy and evaluation measures that further improve prediction accuracy. |
Tasks | Machine Translation |
Published | 2018-05-10 |
URL | http://arxiv.org/abs/1805.04016v2 |
http://arxiv.org/pdf/1805.04016v2.pdf | |
PWC | https://paperswithcode.com/paper/automatic-estimation-of-simultaneous |
Repo | https://github.com/craigastewart/qe_sim_interp |
Framework | none |
Fréchet ChemNet Distance: A metric for generative models for molecules in drug discovery
Title | Fréchet ChemNet Distance: A metric for generative models for molecules in drug discovery |
Authors | Kristina Preuer, Philipp Renz, Thomas Unterthiner, Sepp Hochreiter, Günter Klambauer |
Abstract | The new wave of successful generative models in machine learning has increased the interest in deep learning driven de novo drug design. However, assessing the performance of such generative models is notoriously difficult. Metrics that are typically used to assess the performance of such generative models are the percentage of chemically valid molecules or the similarity to real molecules in terms of particular descriptors, such as the partition coefficient (logP) or druglikeness. However, method comparison is difficult because of the inconsistent use of evaluation metrics, the necessity for multiple metrics, and the fact that some of these measures can easily be tricked by simple rule-based systems. We propose a novel distance measure between two sets of molecules, called Fr'echet ChemNet distance (FCD), that can be used as an evaluation metric for generative models. The FCD is similar to a recently established performance metric for comparing image generation methods, the Fr'echet Inception Distance (FID). Whereas the FID uses one of the hidden layers of InceptionNet, the FCD utilizes the penultimate layer of a deep neural network called ChemNet, which was trained to predict drug activities. Thus, the FCD metric takes into account chemically and biologically relevant information about molecules, and also measures the diversity of the set via the distribution of generated molecules. The FCD’s advantage over previous metrics is that it can detect if generated molecules are a) diverse and have similar b) chemical and c) biological properties as real molecules. We further provide an easy-to-use implementation that only requires the SMILES representation of the generated molecules as input to calculate the FCD. Implementations are available at: https://www.github.com/bioinf-jku/FCD |
Tasks | Drug Discovery |
Published | 2018-03-26 |
URL | http://arxiv.org/abs/1803.09518v3 |
http://arxiv.org/pdf/1803.09518v3.pdf | |
PWC | https://paperswithcode.com/paper/frechet-chemnet-distance-a-metric-for |
Repo | https://github.com/bioinf-jku/FCD |
Framework | tf |
A Neural Multi-sequence Alignment TeCHnique (NeuMATCH)
Title | A Neural Multi-sequence Alignment TeCHnique (NeuMATCH) |
Authors | Pelin Dogan, Boyang Li, Leonid Sigal, Markus Gross |
Abstract | The alignment of heterogeneous sequential data (video to text) is an important and challenging problem. Standard techniques for this task, including Dynamic Time Warping (DTW) and Conditional Random Fields (CRFs), suffer from inherent drawbacks. Mainly, the Markov assumption implies that, given the immediate past, future alignment decisions are independent of further history. The separation between similarity computation and alignment decision also prevents end-to-end training. In this paper, we propose an end-to-end neural architecture where alignment actions are implemented as moving data between stacks of Long Short-term Memory (LSTM) blocks. This flexible architecture supports a large variety of alignment tasks, including one-to-one, one-to-many, skipping unmatched elements, and (with extensions) non-monotonic alignment. Extensive experiments on semi-synthetic and real datasets show that our algorithm outperforms state-of-the-art baselines. |
Tasks | |
Published | 2018-02-19 |
URL | http://arxiv.org/abs/1803.00057v2 |
http://arxiv.org/pdf/1803.00057v2.pdf | |
PWC | https://paperswithcode.com/paper/a-neural-multi-sequence-alignment-technique |
Repo | https://github.com/pelindogan/NeuMATCH |
Framework | none |