January 25, 2020

2766 words 13 mins read

Paper Group NANR 31

Paper Group NANR 31

Adaptive GNN for Image Analysis and Editing. Leveraging Non-Conversational Tasks for Low Resource Slot Filling: Does it help?. Multi-Task Learning of System Dialogue Act Selection for Supervised Pretraining of Goal-Oriented Dialogue Policies. Text-Based Joint Prediction of Numeric and Categorical Attributes of Entities in Knowledge Bases. Atlas of …

Adaptive GNN for Image Analysis and Editing

Title Adaptive GNN for Image Analysis and Editing
Authors Lingyu Liang, Lianwen Jin, Yong Xu
Abstract Graph neural network (GNN) has powerful representation ability, but optimal configurations of GNN are non-trivial to obtain due to diversity of graph structure and cascaded nonlinearities. This paper aims to understand some properties of GNN from a computer vision (CV) perspective. In mathematical analysis, we propose an adaptive GNN model by recursive definition, and derive its relation with two basic operations in CV: filtering and propagation operations. The proposed GNN model is formulated as a label propagation system with guided map, graph Laplacian and node weight. It reveals that 1) the guided map and node weight determine whether a GNN leads to filtering or propagation diffusion, and 2) the kernel of graph Laplacian controls diffusion pattern. In practical verification, we design a new regularization structure with guided feature to produce GNN-based filtering and propagation diffusion to tackle the ill-posed inverse problems of quotient image analysis (QIA), which recovers the reflectance ratio as a signature for image analysis or adjustment. A flexible QIA-GNN framework is constructed to achieve various image-based editing tasks, like face illumination synthesis and low-light image enhancement. Experiments show the effectiveness of the QIA-GNN, and provide new insights of GNN for image analysis and editing.
Tasks Image Enhancement, Low-Light Image Enhancement
Published 2019-12-01
URL http://papers.nips.cc/paper/8622-adaptive-gnn-for-image-analysis-and-editing
PDF http://papers.nips.cc/paper/8622-adaptive-gnn-for-image-analysis-and-editing.pdf
PWC https://paperswithcode.com/paper/adaptive-gnn-for-image-analysis-and-editing
Repo
Framework

Leveraging Non-Conversational Tasks for Low Resource Slot Filling: Does it help?

Title Leveraging Non-Conversational Tasks for Low Resource Slot Filling: Does it help?
Authors Samuel Louvan, Bernardo Magnini
Abstract Slot filling is a core operation for utterance understanding in task-oriented dialogue systems. Slots are typically domain-specific, and adding new domains to a dialogue system involves data and time-intensive processes. A popular technique to address the problem is transfer learning, where it is assumed the availability of a large slot filling dataset for the source domain, to be used to help slot filling on the target domain, with fewer data. In this work, instead, we propose to leverage source tasks based on semantically related non-conversational resources (e.g., semantic sequence tagging datasets), as they are both cheaper to obtain and reusable to several slot filling domains. We show that using auxiliary non-conversational tasks in a multi-task learning setup consistently improves low resource slot filling performance.
Tasks Multi-Task Learning, Slot Filling, Task-Oriented Dialogue Systems, Transfer Learning
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-5911/
PDF https://www.aclweb.org/anthology/W19-5911
PWC https://paperswithcode.com/paper/leveraging-non-conversational-tasks-for-low
Repo
Framework

Multi-Task Learning of System Dialogue Act Selection for Supervised Pretraining of Goal-Oriented Dialogue Policies

Title Multi-Task Learning of System Dialogue Act Selection for Supervised Pretraining of Goal-Oriented Dialogue Policies
Authors Sarah McLeod, Ivana Kruijff-Korbayova, Bernd Kiefer
Abstract This paper describes the use of Multi-Task Neural Networks (NNs) for system dialogue act selection. These models leverage the representations learned by the Natural Language Understanding (NLU) unit to enable robust initialization/bootstrapping of dialogue policies from medium sized initial data sets. We evaluate the models on two goal-oriented dialogue corpora in the travel booking domain. Results show the proposed models improve over models trained without knowledge of NLU tasks.
Tasks Multi-Task Learning
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-5947/
PDF https://www.aclweb.org/anthology/W19-5947
PWC https://paperswithcode.com/paper/multi-task-learning-of-system-dialogue-act
Repo
Framework

Text-Based Joint Prediction of Numeric and Categorical Attributes of Entities in Knowledge Bases

Title Text-Based Joint Prediction of Numeric and Categorical Attributes of Entities in Knowledge Bases
Authors V Thejas, Abhijeet Gupta, Sebastian Pad{'o}
Abstract Collaboratively constructed knowledge bases play an important role in information systems, but are essentially always incomplete. Thus, a large number of models has been developed for Knowledge Base Completion, the task of predicting new attributes of entities given partial descriptions of these entities. Virtually all of these models either concentrate on numeric attributes ({\textless}Italy,GDP,2T{$}{\textgreater}) or they concentrate on categorical attributes ({\textless}Tim Cook,chairman,Apple{\textgreater}). In this paper, we propose a simple feed-forward neural architecture to jointly predict numeric and categorical attributes based on embeddings learned from textual occurrences of the entities in question. Following insights from multi-task learning, our hypothesis is that due to the correlations among attributes of different kinds, joint prediction improves over separate prediction. Our experiments on seven FreeBase domains show that this hypothesis is true of the two attribute types: we find substantial improvements for numeric attributes in the joint model, while performance remains largely unchanged for categorical attributes. Our analysis indicates that this is the case because categorical attributes, many of which describe membership in various classes, provide useful {`}background knowledge{'} for numeric prediction, while this is true to a lesser degree in the inverse direction. |
Tasks Knowledge Base Completion, Multi-Task Learning
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-1137/
PDF https://www.aclweb.org/anthology/R19-1137
PWC https://paperswithcode.com/paper/text-based-joint-prediction-of-numeric-and
Repo
Framework

Atlas of Digital Pathology: A Generalized Hierarchical Histological Tissue Type-Annotated Database for Deep Learning

Title Atlas of Digital Pathology: A Generalized Hierarchical Histological Tissue Type-Annotated Database for Deep Learning
Authors Mahdi S. Hosseini, Lyndon Chan, Gabriel Tse, Michael Tang, Jun Deng, Sajad Norouzi, Corwyn Rowsell, Konstantinos N. Plataniotis, Savvas Damaskinos
Abstract In recent years, computer vision techniques have made large advances in image recognition and been applied to aid radiological diagnosis. Computational pathology aims to develop similar tools for aiding pathologists in diagnosing digitized histopathological slides, which would improve diagnostic accuracy and productivity amidst increasing workloads. However, there is a lack of publicly-available databases of (1) localized patch-level images annotated with (2) a large range of Histological Tissue Type (HTT). As a result, computational pathology research is constrained to diagnosing specific diseases or classifying tissues from specific organs, and cannot be readily generalized to handle unexpected diseases and organs. In this paper, we propose a new digital pathology database, the “Atlas of Digital Pathology” (or ADP), which comprises of 17,668 patch images extracted from 100 slides annotated with up to 57 hierarchical HTTs. Our data is generalized to different tissue types across different organs and aims to provide training data for supervised multi-label learning of patch-level HTT in a digitized whole slide image. We demonstrate the quality of our image labels through pathologist consultation and by training three state-of-the-art neural networks on tissue type classification. Quantitative results support the visually consistency of our data and we demonstrate a tissue type-based visual attention aid as a sample tool that could be developed from our database.
Tasks Multi-Label Learning
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Hosseini_Atlas_of_Digital_Pathology_A_Generalized_Hierarchical_Histological_Tissue_Type-Annotated_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Hosseini_Atlas_of_Digital_Pathology_A_Generalized_Hierarchical_Histological_Tissue_Type-Annotated_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/atlas-of-digital-pathology-a-generalized
Repo
Framework

Progressive Teacher-Student Learning for Early Action Prediction

Title Progressive Teacher-Student Learning for Early Action Prediction
Authors Xionghui Wang, Jian-Fang Hu, Jian-Huang Lai, Jianguo Zhang, Wei-Shi Zheng
Abstract The goal of early action prediction is to recognize actions from partially observed videos with incomplete action executions, which is quite different from action recognition. Predicting early actions is very challenging since the partially observed videos do not contain enough action information for recognition. In this paper, we aim at improving early action prediction by proposing a novel teacher-student learning framework. Our framework involves a teacher model for recognizing actions from full videos, a student model for predicting early actions from partial videos, and a teacher-student learning block for distilling progressive knowledge from teacher to student, crossing different tasks. Extensive experiments on three public action datasets show that the proposed progressive teacher-student learning framework can consistently improve performance of early action prediction model. We have also reported the state-of-the-art performances for early action prediction on all of these sets.
Tasks Temporal Action Localization
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Wang_Progressive_Teacher-Student_Learning_for_Early_Action_Prediction_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Wang_Progressive_Teacher-Student_Learning_for_Early_Action_Prediction_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/progressive-teacher-student-learning-for
Repo
Framework

Dynamic Scene Deblurring With Parameter Selective Sharing and Nested Skip Connections

Title Dynamic Scene Deblurring With Parameter Selective Sharing and Nested Skip Connections
Authors Hongyun Gao, Xin Tao, Xiaoyong Shen, Jiaya Jia
Abstract Dynamic Scene deblurring is a challenging low-level vision task where spatially variant blur is caused by many factors, e.g., camera shake and object motion. Recent study has made significant progress. Compared with the parameter independence scheme [19] and parameter sharing scheme [33], we develop the general principle for constraining the deblurring network structure by proposing the generic and effective selective sharing scheme. Inside the subnetwork of each scale, we propose a nested skip connection structure for the nonlinear transformation modules to replace stacked convolution layers or residual blocks. Besides, we build a new large dataset of blurred/sharp image pairs towards better restoration quality. Comprehensive experimental results show that our parameter selective sharing scheme, nested skip connection structure, and the new dataset are all significant to set a new state-of-the-art in dynamic scene deblurring.
Tasks Deblurring
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Gao_Dynamic_Scene_Deblurring_With_Parameter_Selective_Sharing_and_Nested_Skip_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Gao_Dynamic_Scene_Deblurring_With_Parameter_Selective_Sharing_and_Nested_Skip_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/dynamic-scene-deblurring-with-parameter
Repo
Framework

Coverage of Information Extraction from Sentences and Paragraphs

Title Coverage of Information Extraction from Sentences and Paragraphs
Authors Simon Razniewski, Nitisha Jain, Paramita Mirza, Gerhard Weikum
Abstract Scalar implicatures are language features that imply the negation of stronger statements, e.g., {``}She was married twice{''} typically implicates that she was not married thrice. In this paper we discuss the importance of scalar implicatures in the context of textual information extraction. We investigate how textual features can be used to predict whether a given text segment mentions all objects standing in a certain relationship with a certain subject. Preliminary results on Wikipedia indicate that this prediction is feasible, and yields informative assessments. |
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1583/
PDF https://www.aclweb.org/anthology/D19-1583
PWC https://paperswithcode.com/paper/coverage-of-information-extraction-from
Repo
Framework

A biscriptual morphological transducer for Crimean Tatar

Title A biscriptual morphological transducer for Crimean Tatar
Authors Francis M. Tyers, Jonathan Washington, Darya Kavitskaya, Memduh G{"o}k{\i}rmak, Nick Howell, Remziye Berberova
Abstract
Tasks
Published 2019-02-01
URL https://www.aclweb.org/anthology/W19-6010/
PDF https://www.aclweb.org/anthology/W19-6010
PWC https://paperswithcode.com/paper/a-biscriptual-morphological-transducer-for
Repo
Framework

Finding Sami Cognates with a Character-Based NMT Approach

Title Finding Sami Cognates with a Character-Based NMT Approach
Authors Mika H{"a}m{"a}l{"a}inen, Jack Rueter
Abstract
Tasks
Published 2019-02-01
URL https://www.aclweb.org/anthology/W19-6006/
PDF https://www.aclweb.org/anthology/W19-6006
PWC https://paperswithcode.com/paper/finding-sami-cognates-with-a-character-based
Repo
Framework

Stochastic Filter Groups for Multi-Task CNNs: Learning Specialist and Generalist Convolution Kernels

Title Stochastic Filter Groups for Multi-Task CNNs: Learning Specialist and Generalist Convolution Kernels
Authors Felix J.S. Bragman, Ryutaro Tanno, Sebastien Ourselin, Daniel C. Alexander, Jorge Cardoso
Abstract The performance of multi-task learning in Convolutional Neural Networks (CNNs) hinges on the design of feature sharing between tasks within the architecture. The number of possible sharing patterns are combinatorial in the depth of the network and the number of tasks, and thus hand-crafting an architecture, purely based on the human intuitions of task relationships can be time-consuming and suboptimal. In this paper, we present a probabilistic approach to learning task-specific and shared representations in CNNs for multi-task learning. Specifically, we propose “stochastic filter groups” (SFG), a mechanism to assign convolution kernels in each layer to “specialist” and “generalist” groups, which are specific to and shared across different tasks, respectively. The SFG modules determine the connectivity between layers and the structures of task-specific and shared representations in the network. We employ variational inference to learn the posterior distribution over the possible grouping of kernels and network parameters. Experiments demonstrate the proposed method generalises across multiple tasks and shows improved performance over baseline methods.
Tasks Multi-Task Learning
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Bragman_Stochastic_Filter_Groups_for_Multi-Task_CNNs_Learning_Specialist_and_Generalist_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Bragman_Stochastic_Filter_Groups_for_Multi-Task_CNNs_Learning_Specialist_and_Generalist_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/stochastic-filter-groups-for-multi-task-cnns-1
Repo
Framework

Deep Joint-Semantics Reconstructing Hashing for Large-Scale Unsupervised Cross-Modal Retrieval

Title Deep Joint-Semantics Reconstructing Hashing for Large-Scale Unsupervised Cross-Modal Retrieval
Authors Shupeng Su, Zhisheng Zhong, Chao Zhang
Abstract Cross-modal hashing encodes the multimedia data into a common binary hash space in which the correlations among the samples from different modalities can be effectively measured. Deep cross-modal hashing further improves the retrieval performance as the deep neural networks can generate more semantic relevant features and hash codes. In this paper, we study the unsupervised deep cross-modal hash coding and propose Deep Joint-Semantics Reconstructing Hashing (DJSRH), which has the following two main advantages. First, to learn binary codes that preserve the neighborhood structure of the original data, DJSRH constructs a novel joint-semantics affinity matrix which elaborately integrates the original neighborhood information from different modalities and accordingly is capable to capture the latent intrinsic semantic affinity for the input multi-modal instances. Second, DJSRH later trains the networks to generate binary codes that maximally reconstruct above joint-semantics relations via the proposed reconstructing framework, which is more competent for the batch-wise training as it reconstructs the specific similarity value unlike the common Laplacian constraint merely preserving the similarity order. Extensive experiments demonstrate the significant improvement by DJSRH in various cross-modal retrieval tasks.
Tasks Cross-Modal Retrieval
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Su_Deep_Joint-Semantics_Reconstructing_Hashing_for_Large-Scale_Unsupervised_Cross-Modal_Retrieval_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Su_Deep_Joint-Semantics_Reconstructing_Hashing_for_Large-Scale_Unsupervised_Cross-Modal_Retrieval_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/deep-joint-semantics-reconstructing-hashing
Repo
Framework

``Transforming’’ Delete, Retrieve, Generate Approach for Controlled Text Style Transfer

Title ``Transforming’’ Delete, Retrieve, Generate Approach for Controlled Text Style Transfer |
Authors Akhilesh Sudhakar, Bhargav Upadhyay, Arjun Maheswaran
Abstract Text style transfer is the task of transferring the style of text having certain stylistic attributes, while preserving non-stylistic or content information. In this work we introduce the Generative Style Transformer (GST) - a new approach to rewriting sentences to a target style in the absence of parallel style corpora. GST leverages the power of both, large unsupervised pre-trained language models as well as the Transformer. GST is a part of a larger {`}Delete Retrieve Generate{'} framework, in which we also propose a novel method of deleting style attributes from the source sentence by exploiting the inner workings of the Transformer. Our models outperform state-of-art systems across 5 datasets on sentiment, gender and political slant transfer. We also propose the use of the GLEU metric as an automatic metric of evaluation of style transfer, which we found to compare better with human ratings than the predominantly used BLEU score. |
Tasks Style Transfer, Text Style Transfer
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1322/
PDF https://www.aclweb.org/anthology/D19-1322
PWC https://paperswithcode.com/paper/transforming-delete-retrieve-generate-1
Repo
Framework

Pareto Meets Huber: Efficiently Avoiding Poor Minima in Robust Estimation

Title Pareto Meets Huber: Efficiently Avoiding Poor Minima in Robust Estimation
Authors Christopher Zach, Guillaume Bourmaud
Abstract Robust cost optimization is the task of fitting parameters to data points containing outliers. In particular, we focus on large-scale computer vision problems, such as bundle adjustment, where Non-Linear Least Square (NLLS) solvers are the current workhorse. In this context, NLLS-based state of the art algorithms have been designed either to quickly improve the target objective and find a local minimum close to the initial value of the parameters, or to have a strong ability to escape poor local minima. In this paper, we propose a novel algorithm relying on multi-objective optimization which allows to match those two properties. We experimentally demonstrate that our algorithm has an ability to escape poor local minima that is on par with the best performing algorithms with a faster decrease of the target objective.
Tasks
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Zach_Pareto_Meets_Huber_Efficiently_Avoiding_Poor_Minima_in_Robust_Estimation_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Zach_Pareto_Meets_Huber_Efficiently_Avoiding_Poor_Minima_in_Robust_Estimation_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/pareto-meets-huber-efficiently-avoiding-poor
Repo
Framework

Deep Bayesian Natural Language Processing

Title Deep Bayesian Natural Language Processing
Authors Jen-Tzung Chien
Abstract This introductory tutorial addresses the advances in deep Bayesian learning for natural language with ubiquitous applications ranging from speech recognition to document summarization, text classification, text segmentation, information extraction, image caption generation, sentence generation, dialogue control, sentiment classification, recommendation system, question answering and machine translation, to name a few. Traditionally, {}deep learning{''} is taken to be a learning process where the inference or optimization is based on the real-valued deterministic model. The {}semantic structure{''} in words, sentences, entities, actions and documents drawn from a large vocabulary may not be well expressed or correctly optimized in mathematical logic or computer programs. The {``}distribution function{''} in discrete or continuous latent variable model for natural language may not be properly decomposed or estimated. This tutorial addresses the fundamentals of statistical models and neural networks, and focus on a series of advanced Bayesian models and deep models including hierarchical Dirichlet process, Chinese restaurant process, hierarchical Pitman-Yor process, Indian buffet process, recurrent neural network, long short-term memory, sequence-to-sequence model, variational auto-encoder, generative adversarial network, attention mechanism, memory-augmented neural network, skip neural network, stochastic neural network, predictive state neural network and policy neural network. We present how these models are connected and why they work for a variety of applications on symbolic and complex patterns in natural language. The variational inference and sampling method are formulated to tackle the optimization for complicated models. The word and sentence embeddings, clustering and co-clustering are merged with linguistic and semantic constraints. A series of case studies and domain applications are presented to tackle different issues in deep Bayesian processing, learning and understanding. At last, we will point out a number of directions and outlooks for future studies. |
Tasks Document Summarization, Machine Translation, Question Answering, Sentence Embeddings, Sentiment Analysis, Speech Recognition, Text Classification
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-4006/
PDF https://www.aclweb.org/anthology/P19-4006
PWC https://paperswithcode.com/paper/deep-bayesian-natural-language-processing
Repo
Framework
comments powered by Disqus