Paper Group ANR 724
Multi-Task Learning for Speaker-Role Adaptation in Neural Conversation Models. Ensembling Factored Neural Machine Translation Models for Automatic Post-Editing and Quality Estimation. Deep Learning for Computational Chemistry. Generalized Zero-Shot Learning for Action Recognition with Web-Scale Video Data. Fine-grained Visual-textual Representation …
Multi-Task Learning for Speaker-Role Adaptation in Neural Conversation Models
Title | Multi-Task Learning for Speaker-Role Adaptation in Neural Conversation Models |
Authors | Yi Luan, Chris Brockett, Bill Dolan, Jianfeng Gao, Michel Galley |
Abstract | Building a persona-based conversation agent is challenging owing to the lack of large amounts of speaker-specific conversation data for model training. This paper addresses the problem by proposing a multi-task learning approach to training neural conversation models that leverages both conversation data across speakers and other types of data pertaining to the speaker and speaker roles to be modeled. Experiments show that our approach leads to significant improvements over baseline model quality, generating responses that capture more precisely speakers’ traits and speaking styles. The model offers the benefits of being algorithmically simple and easy to implement, and not relying on large quantities of data representing specific individual speakers. |
Tasks | Multi-Task Learning |
Published | 2017-10-20 |
URL | http://arxiv.org/abs/1710.07388v1 |
http://arxiv.org/pdf/1710.07388v1.pdf | |
PWC | https://paperswithcode.com/paper/multi-task-learning-for-speaker-role |
Repo | |
Framework | |
Ensembling Factored Neural Machine Translation Models for Automatic Post-Editing and Quality Estimation
Title | Ensembling Factored Neural Machine Translation Models for Automatic Post-Editing and Quality Estimation |
Authors | Chris Hokamp |
Abstract | This work presents a novel approach to Automatic Post-Editing (APE) and Word-Level Quality Estimation (QE) using ensembles of specialized Neural Machine Translation (NMT) systems. Word-level features that have proven effective for QE are included as input factors, expanding the representation of the original source and the machine translation hypothesis, which are used to generate an automatically post-edited hypothesis. We train a suite of NMT models that use different input representations, but share the same output space. These models are then ensembled together, and tuned for both the APE and the QE task. We thus attempt to connect the state-of-the-art approaches to APE and QE within a single framework. Our models achieve state-of-the-art results in both tasks, with the only difference in the tuning step which learns weights for each component of the ensemble. |
Tasks | Automatic Post-Editing, Machine Translation |
Published | 2017-06-15 |
URL | http://arxiv.org/abs/1706.05083v2 |
http://arxiv.org/pdf/1706.05083v2.pdf | |
PWC | https://paperswithcode.com/paper/ensembling-factored-neural-machine |
Repo | |
Framework | |
Deep Learning for Computational Chemistry
Title | Deep Learning for Computational Chemistry |
Authors | Garrett B. Goh, Nathan O. Hodas, Abhinav Vishnu |
Abstract | The rise and fall of artificial neural networks is well documented in the scientific literature of both computer science and computational chemistry. Yet almost two decades later, we are now seeing a resurgence of interest in deep learning, a machine learning algorithm based on multilayer neural networks. Within the last few years, we have seen the transformative impact of deep learning in many domains, particularly in speech recognition and computer vision, to the extent that the majority of expert practitioners in those field are now regularly eschewing prior established models in favor of deep learning models. In this review, we provide an introductory overview into the theory of deep neural networks and their unique properties that distinguish them from traditional machine learning algorithms used in cheminformatics. By providing an overview of the variety of emerging applications of deep neural networks, we highlight its ubiquity and broad applicability to a wide range of challenges in the field, including QSAR, virtual screening, protein structure prediction, quantum chemistry, materials design and property prediction. In reviewing the performance of deep neural networks, we observed a consistent outperformance against non-neural networks state-of-the-art models across disparate research topics, and deep neural network based models often exceeded the “glass ceiling” expectations of their respective tasks. Coupled with the maturity of GPU-accelerated computing for training deep neural networks and the exponential growth of chemical data on which to train these networks on, we anticipate that deep learning algorithms will be a valuable tool for computational chemistry. |
Tasks | Speech Recognition |
Published | 2017-01-17 |
URL | http://arxiv.org/abs/1701.04503v1 |
http://arxiv.org/pdf/1701.04503v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-learning-for-computational-chemistry |
Repo | |
Framework | |
Generalized Zero-Shot Learning for Action Recognition with Web-Scale Video Data
Title | Generalized Zero-Shot Learning for Action Recognition with Web-Scale Video Data |
Authors | Kun Liu, Wu Liu, Huadong Ma, Wenbing Huang, Xiongxiong Dong |
Abstract | Action recognition in surveillance video makes our life safer by detecting the criminal events or predicting violent emergencies. However, efficient action recognition is not free of difficulty. First, there are so many action classes in daily life that we cannot pre-define all possible action classes beforehand. Moreover, it is very hard to collect real-word videos for certain particular actions such as steal and street fight due to legal restrictions and privacy protection. These challenges make existing data-driven recognition methods insufficient to attain desired performance. Zero-shot learning is potential to be applied to solve these issues since it can perform classification without positive example. Nevertheless, current zero-shot learning algorithms have been studied under the unreasonable setting where seen classes are absent during the testing phase. Motivated by this, we study the task of action recognition in surveillance video under a more realistic \emph{generalized zero-shot setting}, where testing data contains both seen and unseen classes. To our best knowledge, this is the first work to study video action recognition under the generalized zero-shot setting. We firstly perform extensive empirical studies on several existing zero-shot leaning approaches under this new setting on a web-scale video data. Our experimental results demonstrate that, under the generalize setting, typical zero-shot learning methods are no longer effective for the dataset we applied. Then, we propose a method for action recognition by deploying generalized zero-shot learning, which transfers the knowledge of web video to detect the anomalous actions in surveillance videos. To verify the effectiveness of our proposed method, we further construct a new surveillance video dataset consisting of nine action classes related to the public safety situation. |
Tasks | Temporal Action Localization, Zero-Shot Learning |
Published | 2017-10-20 |
URL | http://arxiv.org/abs/1710.07455v1 |
http://arxiv.org/pdf/1710.07455v1.pdf | |
PWC | https://paperswithcode.com/paper/generalized-zero-shot-learning-for-action |
Repo | |
Framework | |
Fine-grained Visual-textual Representation Learning
Title | Fine-grained Visual-textual Representation Learning |
Authors | Xiangteng He, Yuxin Peng |
Abstract | Fine-grained visual categorization is to recognize hundreds of subcategories belonging to the same basic-level category, which is a highly challenging task due to the quite subtle and local visual distinctions among similar subcategories. Most existing methods generally learn part detectors to discover discriminative regions for better categorization performance. However, not all parts are beneficial and indispensable for visual categorization, and the setting of part detector number heavily relies on prior knowledge as well as experimental validation. As is known to all, when we describe the object of an image via textual descriptions, we mainly focus on the pivotal characteristics, and rarely pay attention to common characteristics as well as the background areas. This is an involuntary transfer from human visual attention to textual attention, which leads to the fact that textual attention tells us how many and which parts are discriminative and significant to categorization. So textual attention could help us to discover visual attention in image. Inspired by this, we propose a fine-grained visual-textual representation learning (VTRL) approach, and its main contributions are: (1) Fine-grained visual-textual pattern mining devotes to discovering discriminative visual-textual pairwise information for boosting categorization performance through jointly modeling vision and text with generative adversarial networks (GANs), which automatically and adaptively discovers discriminative parts. (2) Visual-textual representation learning jointly combines visual and textual information, which preserves the intra-modality and inter-modality information to generate complementary fine-grained representation, as well as further improves categorization performance. |
Tasks | Fine-Grained Visual Categorization, Representation Learning |
Published | 2017-08-31 |
URL | http://arxiv.org/abs/1709.00340v4 |
http://arxiv.org/pdf/1709.00340v4.pdf | |
PWC | https://paperswithcode.com/paper/fine-grained-visual-textual-representation |
Repo | |
Framework | |
Exploring the Imposition of Synaptic Precision Restrictions For Evolutionary Synthesis of Deep Neural Networks
Title | Exploring the Imposition of Synaptic Precision Restrictions For Evolutionary Synthesis of Deep Neural Networks |
Authors | Mohammad Javad Shafiee, Francis Li, Alexander Wong |
Abstract | A key contributing factor to incredible success of deep neural networks has been the significant rise on massively parallel computing devices allowing researchers to greatly increase the size and depth of deep neural networks, leading to significant improvements in modeling accuracy. Although deeper, larger, or complex deep neural networks have shown considerable promise, the computational complexity of such networks is a major barrier to utilization in resource-starved scenarios. We explore the synaptogenesis of deep neural networks in the formation of efficient deep neural network architectures within an evolutionary deep intelligence framework, where a probabilistic generative modeling strategy is introduced to stochastically synthesize increasingly efficient yet effective offspring deep neural networks over generations, mimicking evolutionary processes such as heredity, random mutation, and natural selection in a probabilistic manner. In this study, we primarily explore the imposition of synaptic precision restrictions and its impact on the evolutionary synthesis of deep neural networks to synthesize more efficient network architectures tailored for resource-starved scenarios. Experimental results show significant improvements in synaptic efficiency (~10X decrease for GoogLeNet-based DetectNet) and inference speed (>5X increase for GoogLeNet-based DetectNet) while preserving modeling accuracy. |
Tasks | |
Published | 2017-07-01 |
URL | http://arxiv.org/abs/1707.00095v1 |
http://arxiv.org/pdf/1707.00095v1.pdf | |
PWC | https://paperswithcode.com/paper/exploring-the-imposition-of-synaptic |
Repo | |
Framework | |
One-Shot Fine-Grained Instance Retrieval
Title | One-Shot Fine-Grained Instance Retrieval |
Authors | Hantao Yao, Shiliang Zhang, Yongdong Zhang, Jintao Li, Qi Tian |
Abstract | Fine-Grained Visual Categorization (FGVC) has achieved significant progress recently. However, the number of fine-grained species could be huge and dynamically increasing in real scenarios, making it difficult to recognize unseen objects under the current FGVC framework. This raises an open issue to perform large-scale fine-grained identification without a complete training set. Aiming to conquer this issue, we propose a retrieval task named One-Shot Fine-Grained Instance Retrieval (OSFGIR). “One-Shot” denotes the ability of identifying unseen objects through a fine-grained retrieval task assisted with an incomplete auxiliary training set. This paper first presents the detailed description to OSFGIR task and our collected OSFGIR-378K dataset. Next, we propose the Convolutional and Normalization Networks (CN-Nets) learned on the auxiliary dataset to generate a concise and discriminative representation. Finally, we present a coarse-to-fine retrieval framework consisting of three components, i.e., coarse retrieval, fine-grained retrieval, and query expansion, respectively. The framework progressively retrieves images with similar semantics, and performs fine-grained identification. Experiments show our OSFGIR framework achieves significantly better accuracy and efficiency than existing FGVC and image retrieval methods, thus could be a better solution for large-scale fine-grained object identification. |
Tasks | Fine-Grained Visual Categorization, Image Retrieval |
Published | 2017-07-04 |
URL | http://arxiv.org/abs/1707.00811v1 |
http://arxiv.org/pdf/1707.00811v1.pdf | |
PWC | https://paperswithcode.com/paper/one-shot-fine-grained-instance-retrieval |
Repo | |
Framework | |
Exploiting Active Subspaces in Global Optimization: How Complex is your Problem?
Title | Exploiting Active Subspaces in Global Optimization: How Complex is your Problem? |
Authors | Pramudita Satria Palar, Koji Shimoyama |
Abstract | When applying optimization method to a real-world problem, the possession of prior knowledge and preliminary analysis on the landscape of a global optimization problem can give us an insight into the complexity of the problem. This knowledge can better inform us in deciding what optimization method should be used to tackle the problem. However, this analysis becomes problematic when the dimensionality of the problem is high. This paper presents a framework to take a deeper look at the global optimization problem to be tackled: by analyzing the low-dimensional representation of the problem through discovering the active subspaces of the given problem. The virtue of this is that the problem’s complexity can be visualized in a one or two-dimensional plot, thus allow one to get a better grip about the problem’s difficulty. One could then have a better idea regarding the complexity of their problem to determine the choice of global optimizer or what surrogate-model type to be used. Furthermore, we also demonstrate how the active subspaces can be used to perform design exploration and analysis. |
Tasks | |
Published | 2017-07-09 |
URL | http://arxiv.org/abs/1707.02533v1 |
http://arxiv.org/pdf/1707.02533v1.pdf | |
PWC | https://paperswithcode.com/paper/exploiting-active-subspaces-in-global |
Repo | |
Framework | |
Context encoding enables machine learning-based quantitative photoacoustics
Title | Context encoding enables machine learning-based quantitative photoacoustics |
Authors | Thomas Kirchner, Janek Gröhl, Lena Maier-Hein |
Abstract | Real-time monitoring of functional tissue parameters, such as local blood oxygenation, based on optical imaging could provide groundbreaking advances in the diagnosis and interventional therapy of various diseases. While photoacoustic (PA) imaging is a novel modality with great potential to measure optical absorption deep inside tissue, quantification of the measurements remains a major challenge. In this paper, we introduce the first machine learning based approach to quantitative PA imaging (qPAI), which relies on learning the fluence in a voxel to deduce the corresponding optical absorption. The method encodes relevant information of the measured signal and the characteristics of the imaging system in voxel-based feature vectors, which allow the generation of thousands of training samples from a single simulated PA image. Comprehensive in silico experiments suggest that context encoding (CE)-qPAI enables highly accurate and robust quantification of the local fluence and thereby the optical absorption from PA images. |
Tasks | |
Published | 2017-06-12 |
URL | http://arxiv.org/abs/1706.03595v2 |
http://arxiv.org/pdf/1706.03595v2.pdf | |
PWC | https://paperswithcode.com/paper/context-encoding-enables-machine-learning |
Repo | |
Framework | |
Order-Preserving Abstractive Summarization for Spoken Content Based on Connectionist Temporal Classification
Title | Order-Preserving Abstractive Summarization for Spoken Content Based on Connectionist Temporal Classification |
Authors | Bo-Ru Lu, Frank Shyu, Yun-Nung Chen, Hung-Yi Lee, Lin-shan Lee |
Abstract | Connectionist temporal classification (CTC) is a powerful approach for sequence-to-sequence learning, and has been popularly used in speech recognition. The central ideas of CTC include adding a label “blank” during training. With this mechanism, CTC eliminates the need of segment alignment, and hence has been applied to various sequence-to-sequence learning problems. In this work, we applied CTC to abstractive summarization for spoken content. The “blank” in this case implies the corresponding input data are less important or noisy; thus it can be ignored. This approach was shown to outperform the existing methods in term of ROUGE scores over Chinese Gigaword and MATBN corpora. This approach also has the nice property that the ordering of words or characters in the input documents can be better preserved in the generated summaries. |
Tasks | Abstractive Text Summarization, Speech Recognition |
Published | 2017-09-16 |
URL | http://arxiv.org/abs/1709.05475v2 |
http://arxiv.org/pdf/1709.05475v2.pdf | |
PWC | https://paperswithcode.com/paper/order-preserving-abstractive-summarization |
Repo | |
Framework | |
Convergent Tree Backup and Retrace with Function Approximation
Title | Convergent Tree Backup and Retrace with Function Approximation |
Authors | Ahmed Touati, Pierre-Luc Bacon, Doina Precup, Pascal Vincent |
Abstract | Off-policy learning is key to scaling up reinforcement learning as it allows to learn about a target policy from the experience generated by a different behavior policy. Unfortunately, it has been challenging to combine off-policy learning with function approximation and multi-step bootstrapping in a way that leads to both stable and efficient algorithms. In this work, we show that the \textsc{Tree Backup} and \textsc{Retrace} algorithms are unstable with linear function approximation, both in theory and in practice with specific examples. Based on our analysis, we then derive stable and efficient gradient-based algorithms using a quadratic convex-concave saddle-point formulation. By exploiting the problem structure proper to these algorithms, we are able to provide convergence guarantees and finite-sample bounds. The applicability of our new analysis also goes beyond \textsc{Tree Backup} and \textsc{Retrace} and allows us to provide new convergence rates for the GTD and GTD2 algorithms without having recourse to projections or Polyak averaging. |
Tasks | |
Published | 2017-05-25 |
URL | http://arxiv.org/abs/1705.09322v4 |
http://arxiv.org/pdf/1705.09322v4.pdf | |
PWC | https://paperswithcode.com/paper/convergent-tree-backup-and-retrace-with |
Repo | |
Framework | |
Proxy Non-Discrimination in Data-Driven Systems
Title | Proxy Non-Discrimination in Data-Driven Systems |
Authors | Anupam Datta, Matt Fredrikson, Gihyuk Ko, Piotr Mardziel, Shayak Sen |
Abstract | Machine learnt systems inherit biases against protected classes, historically disparaged groups, from training data. Usually, these biases are not explicit, they rely on subtle correlations discovered by training algorithms, and are therefore difficult to detect. We formalize proxy discrimination in data-driven systems, a class of properties indicative of bias, as the presence of protected class correlates that have causal influence on the system’s output. We evaluate an implementation on a corpus of social datasets, demonstrating how to validate systems against these properties and to repair violations where they occur. |
Tasks | |
Published | 2017-07-25 |
URL | http://arxiv.org/abs/1707.08120v1 |
http://arxiv.org/pdf/1707.08120v1.pdf | |
PWC | https://paperswithcode.com/paper/proxy-non-discrimination-in-data-driven |
Repo | |
Framework | |
Effective scaling registration approach by imposing the emphasis on the scale factor
Title | Effective scaling registration approach by imposing the emphasis on the scale factor |
Authors | Minmin Xu, Siyu Xu, Jihua Zhu, Yaochen Li, Jun Wang, Huimin Lu |
Abstract | This paper proposes an effective approach for the scaling registration of $m$-D point sets. Different from the rigid transformation, the scaling registration can not be formulated into the common least square function due to the ill-posed problem caused by the scale factor. Therefore, this paper designs a novel objective function for the scaling registration problem. The appearance of this objective function is a rational fraction, where the numerator item is the least square error and the denominator item is the square of the scale factor. By imposing the emphasis on scale factor, the ill-posed problem can be avoided in the scaling registration. Subsequently, the new objective function can be solved by the proposed scaling iterative closest point (ICP) algorithm, which can obtain the optimal scaling transformation. For the practical applications, the scaling ICP algorithm is further extended to align partially overlapping point sets. Finally, the proposed approach is tested on public data sets and applied to merging grid maps of different resolutions. Experimental results demonstrate its superiority over previous approaches on efficiency and robustness. |
Tasks | |
Published | 2017-04-28 |
URL | http://arxiv.org/abs/1705.00086v2 |
http://arxiv.org/pdf/1705.00086v2.pdf | |
PWC | https://paperswithcode.com/paper/effective-scaling-registration-approach-by |
Repo | |
Framework | |
Generalization Bounds of SGLD for Non-convex Learning: Two Theoretical Viewpoints
Title | Generalization Bounds of SGLD for Non-convex Learning: Two Theoretical Viewpoints |
Authors | Wenlong Mou, Liwei Wang, Xiyu Zhai, Kai Zheng |
Abstract | Algorithm-dependent generalization error bounds are central to statistical learning theory. A learning algorithm may use a large hypothesis space, but the limited number of iterations controls its model capacity and generalization error. The impacts of stochastic gradient methods on generalization error for non-convex learning problems not only have important theoretical consequences, but are also critical to generalization errors of deep learning. In this paper, we study the generalization errors of Stochastic Gradient Langevin Dynamics (SGLD) with non-convex objectives. Two theories are proposed with non-asymptotic discrete-time analysis, using Stability and PAC-Bayesian results respectively. The stability-based theory obtains a bound of $O\left(\frac{1}{n}L\sqrt{\beta T_k}\right)$, where $L$ is uniform Lipschitz parameter, $\beta$ is inverse temperature, and $T_k$ is aggregated step sizes. For PAC-Bayesian theory, though the bound has a slower $O(1/\sqrt{n})$ rate, the contribution of each step is shown with an exponentially decaying factor by imposing $\ell^2$ regularization, and the uniform Lipschitz constant is also replaced by actual norms of gradients along trajectory. Our bounds have no implicit dependence on dimensions, norms or other capacity measures of parameter, which elegantly characterizes the phenomenon of “Fast Training Guarantees Generalization” in non-convex settings. This is the first algorithm-dependent result with reasonable dependence on aggregated step sizes for non-convex learning, and has important implications to statistical learning aspects of stochastic gradient methods in complicated models such as deep learning. |
Tasks | |
Published | 2017-07-19 |
URL | http://arxiv.org/abs/1707.05947v1 |
http://arxiv.org/pdf/1707.05947v1.pdf | |
PWC | https://paperswithcode.com/paper/generalization-bounds-of-sgld-for-non-convex |
Repo | |
Framework | |
Residual Expansion Algorithm: Fast and Effective Optimization for Nonconvex Least Squares Problems
Title | Residual Expansion Algorithm: Fast and Effective Optimization for Nonconvex Least Squares Problems |
Authors | Daiki Ikami, Toshihiko Yamasaki, Kiyoharu Aizawa |
Abstract | We propose the residual expansion (RE) algorithm: a global (or near-global) optimization method for nonconvex least squares problems. Unlike most existing nonconvex optimization techniques, the RE algorithm is not based on either stochastic or multi-point searches; therefore, it can achieve fast global optimization. Moreover, the RE algorithm is easy to implement and successful in high-dimensional optimization. The RE algorithm exhibits excellent empirical performance in terms of k-means clustering, point-set registration, optimized product quantization, and blind image deblurring. |
Tasks | Blind Image Deblurring, Deblurring, Quantization |
Published | 2017-05-26 |
URL | http://arxiv.org/abs/1705.09549v1 |
http://arxiv.org/pdf/1705.09549v1.pdf | |
PWC | https://paperswithcode.com/paper/residual-expansion-algorithm-fast-and |
Repo | |
Framework | |