Paper Group AWR 122
ARMIN: Towards a More Efficient and Light-weight Recurrent Memory Network. Unsupervised Attributed Multiplex Network Embedding. Barack’s Wife Hillary: Using Knowledge-Graphs for Fact-Aware Language Modeling. 2D and 3D Segmentation of uncertain local collagen fiber orientations in SHG microscopy. HiGitClass: Keyword-Driven Hierarchical Classificatio …
ARMIN: Towards a More Efficient and Light-weight Recurrent Memory Network
Title | ARMIN: Towards a More Efficient and Light-weight Recurrent Memory Network |
Authors | Zhangheng Li, Jia-Xing Zhong, Jingjia Huang, Tao Zhang, Thomas Li, Ge Li |
Abstract | In recent years, memory-augmented neural networks(MANNs) have shown promising power to enhance the memory ability of neural networks for sequential processing tasks. However, previous MANNs suffer from complex memory addressing mechanism, making them relatively hard to train and causing computational overheads. Moreover, many of them reuse the classical RNN structure such as LSTM for memory processing, causing inefficient exploitations of memory information. In this paper, we introduce a novel MANN, the Auto-addressing and Recurrent Memory Integrating Network (ARMIN) to address these issues. The ARMIN only utilizes hidden state ht for automatic memory addressing, and uses a novel RNN cell for refined integration of memory information. Empirical results on a variety of experiments demonstrate that the ARMIN is more light-weight and efficient compared to existing memory networks. Moreover, we demonstrate that the ARMIN can achieve much lower computational overhead than vanilla LSTM while keeping similar performances. Codes are available on github.com/zoharli/armin. |
Tasks | |
Published | 2019-06-28 |
URL | https://arxiv.org/abs/1906.12087v1 |
https://arxiv.org/pdf/1906.12087v1.pdf | |
PWC | https://paperswithcode.com/paper/armin-towards-a-more-efficient-and-light |
Repo | https://github.com/zoharli/armin |
Framework | pytorch |
Unsupervised Attributed Multiplex Network Embedding
Title | Unsupervised Attributed Multiplex Network Embedding |
Authors | Chanyoung Park, Donghyun Kim, Jiawei Han, Hwanjo Yu |
Abstract | Nodes in a multiplex network are connected by multiple types of relations. However, most existing network embedding methods assume that only a single type of relation exists between nodes. Even for those that consider the multiplexity of a network, they overlook node attributes, resort to node labels for training, and fail to model the global properties of a graph. We present a simple yet effective unsupervised network embedding method for attributed multiplex network called DMGI, inspired by Deep Graph Infomax (DGI) that maximizes the mutual information between local patches of a graph, and the global representation of the entire graph. We devise a systematic way to jointly integrate the node embeddings from multiple graphs by introducing 1) the consensus regularization framework that minimizes the disagreements among the relation-type specific node embeddings, and 2) the universal discriminator that discriminates true samples regardless of the relation types. We also show that the attention mechanism infers the importance of each relation type, and thus can be useful for filtering unnecessary relation types as a preprocessing step. Extensive experiments on various downstream tasks demonstrate that DMGI outperforms the state-of-the-art methods, even though DMGI is fully unsupervised. |
Tasks | Network Embedding |
Published | 2019-11-15 |
URL | https://arxiv.org/abs/1911.06750v2 |
https://arxiv.org/pdf/1911.06750v2.pdf | |
PWC | https://paperswithcode.com/paper/unsupervised-attributed-multiplex-network |
Repo | https://github.com/pcy1302/DMGI |
Framework | pytorch |
Barack’s Wife Hillary: Using Knowledge-Graphs for Fact-Aware Language Modeling
Title | Barack’s Wife Hillary: Using Knowledge-Graphs for Fact-Aware Language Modeling |
Authors | Robert L. Logan IV, Nelson F. Liu, Matthew E. Peters, Matt Gardner, Sameer Singh |
Abstract | Modeling human language requires the ability to not only generate fluent text but also encode factual knowledge. However, traditional language models are only capable of remembering facts seen at training time, and often have difficulty recalling them. To address this, we introduce the knowledge graph language model (KGLM), a neural language model with mechanisms for selecting and copying facts from a knowledge graph that are relevant to the context. These mechanisms enable the model to render information it has never seen before, as well as generate out-of-vocabulary tokens. We also introduce the Linked WikiText-2 dataset, a corpus of annotated text aligned to the Wikidata knowledge graph whose contents (roughly) match the popular WikiText-2 benchmark. In experiments, we demonstrate that the KGLM achieves significantly better performance than a strong baseline language model. We additionally compare different language model’s ability to complete sentences requiring factual knowledge, showing that the KGLM outperforms even very large language models in generating facts. |
Tasks | Knowledge Graphs, Language Modelling |
Published | 2019-06-17 |
URL | https://arxiv.org/abs/1906.07241v2 |
https://arxiv.org/pdf/1906.07241v2.pdf | |
PWC | https://paperswithcode.com/paper/baracks-wife-hillary-using-knowledge-graphs |
Repo | https://github.com/rloganiv/kglm-model |
Framework | pytorch |
2D and 3D Segmentation of uncertain local collagen fiber orientations in SHG microscopy
Title | 2D and 3D Segmentation of uncertain local collagen fiber orientations in SHG microscopy |
Authors | Lars Schmarje, Claudius Zelenka, Ulf Geisen, Claus-C. Glüer, Reinhard Koch |
Abstract | Collagen fiber orientations in bones, visible with Second Harmonic Generation (SHG) microscopy, represent the inner structure and its alteration due to influences like cancer. While analyses of these orientations are valuable for medical research, it is not feasible to analyze the needed large amounts of local orientations manually. Since we have uncertain borders for these local orientations only rough regions can be segmented instead of a pixel-wise segmentation. We analyze the effect of these uncertain borders on human performance by a user study. Furthermore, we compare a variety of 2D and 3D methods such as classical approaches like Fourier analysis with state-of-the-art deep neural networks for the classification of local fiber orientations. We present a general way to use pretrained 2D weights in 3D neural networks, such as Inception-ResNet-3D a 3D extension of Inception-ResNet-v2. In a 10 fold cross-validation our two stage segmentation based on Inception-ResNet-3D and transferred 2D ImageNet weights achieves a human comparable accuracy. |
Tasks | |
Published | 2019-07-30 |
URL | https://arxiv.org/abs/1907.12868v1 |
https://arxiv.org/pdf/1907.12868v1.pdf | |
PWC | https://paperswithcode.com/paper/2d-and-3d-segmentation-of-uncertain-local |
Repo | https://github.com/Emprime/uncertain-fiber-segmentation |
Framework | tf |
HiGitClass: Keyword-Driven Hierarchical Classification of GitHub Repositories
Title | HiGitClass: Keyword-Driven Hierarchical Classification of GitHub Repositories |
Authors | Yu Zhang, Frank F. Xu, Sha Li, Yu Meng, Xuan Wang, Qi Li, Jiawei Han |
Abstract | GitHub has become an important platform for code sharing and scientific exchange. With the massive number of repositories available, there is a pressing need for topic-based search. Even though the topic label functionality has been introduced, the majority of GitHub repositories do not have any labels, impeding the utility of search and topic-based analysis. This work targets the automatic repository classification problem as \textit{keyword-driven hierarchical classification}. Specifically, users only need to provide a label hierarchy with keywords to supply as supervision. This setting is flexible, adaptive to the users’ needs, accounts for the different granularity of topic labels and requires minimal human effort. We identify three key challenges of this problem, namely (1) the presence of multi-modal signals; (2) supervision scarcity and bias; (3) supervision format mismatch. In recognition of these challenges, we propose the \textsc{HiGitClass} framework, comprising of three modules: heterogeneous information network embedding; keyword enrichment; topic modeling and pseudo document generation. Experimental results on two GitHub repository collections confirm that \textsc{HiGitClass} is superior to existing weakly-supervised and dataless hierarchical classification methods, especially in its ability to integrate both structured and unstructured data for repository classification. |
Tasks | Network Embedding |
Published | 2019-10-16 |
URL | https://arxiv.org/abs/1910.07115v1 |
https://arxiv.org/pdf/1910.07115v1.pdf | |
PWC | https://paperswithcode.com/paper/higitclass-keyword-driven-hierarchical |
Repo | https://github.com/yuzhimanhua/HiGitClass |
Framework | pytorch |
Learning Attention-based Embeddings for Relation Prediction in Knowledge Graphs
Title | Learning Attention-based Embeddings for Relation Prediction in Knowledge Graphs |
Authors | Deepak Nathani, Jatin Chauhan, Charu Sharma, Manohar Kaul |
Abstract | The recent proliferation of knowledge graphs (KGs) coupled with incomplete or partial information, in the form of missing relations (links) between entities, has fueled a lot of research on knowledge base completion (also known as relation prediction). Several recent works suggest that convolutional neural network (CNN) based models generate richer and more expressive feature embeddings and hence also perform well on relation prediction. However, we observe that these KG embeddings treat triples independently and thus fail to cover the complex and hidden information that is inherently implicit in the local neighborhood surrounding a triple. To this effect, our paper proposes a novel attention based feature embedding that captures both entity and relation features in any given entity’s neighborhood. Additionally, we also encapsulate relation clusters and multihop relations in our model. Our empirical study offers insights into the efficacy of our attention based model and we show marked performance gains in comparison to state of the art methods on all datasets. |
Tasks | Knowledge Base Completion, Knowledge Graph Embeddings, Knowledge Graphs, Link Prediction |
Published | 2019-06-04 |
URL | https://arxiv.org/abs/1906.01195v1 |
https://arxiv.org/pdf/1906.01195v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-attention-based-embeddings-for |
Repo | https://github.com/deepakn97/relationPrediction |
Framework | pytorch |
Active Learning with Gaussian Processes for High Throughput Phenotyping
Title | Active Learning with Gaussian Processes for High Throughput Phenotyping |
Authors | Sumit Kumar, Wenhao Luo, George Kantor, Katia Sycara |
Abstract | A looming question that must be solved before robotic plant phenotyping capabilities can have significant impact to crop improvement programs is scalability. High Throughput Phenotyping (HTP) uses robotic technologies to analyze crops in order to determine species with favorable traits, however, the current practices rely on exhaustive coverage and data collection from the entire crop field being monitored under the breeding experiment. This works well in relatively small agricultural fields but can not be scaled to the larger ones, thus limiting the progress of genetics research. In this work, we propose an active learning algorithm to enable an autonomous system to collect the most informative samples in order to accurately learn the distribution of phenotypes in the field with the help of a Gaussian Process model. We demonstrate the superior performance of our proposed algorithm compared to the current practices on sorghum phenotype data collection. |
Tasks | Active Learning, Gaussian Processes |
Published | 2019-01-21 |
URL | http://arxiv.org/abs/1901.06803v1 |
http://arxiv.org/pdf/1901.06803v1.pdf | |
PWC | https://paperswithcode.com/paper/active-learning-with-gaussian-processes-for |
Repo | https://github.com/sumitsk/algp |
Framework | pytorch |
HumanMeshNet: Polygonal Mesh Recovery of Humans
Title | HumanMeshNet: Polygonal Mesh Recovery of Humans |
Authors | Abbhinav Venkat, Chaitanya Patel, Yudhik Agrawal, Avinash Sharma |
Abstract | 3D Human Body Reconstruction from a monocular image is an important problem in computer vision with applications in virtual and augmented reality platforms, animation industry, en-commerce domain, etc. While several of the existing works formulate it as a volumetric or parametric learning with complex and indirect reliance on re-projections of the mesh, we would like to focus on implicitly learning the mesh representation. To that end, we propose a novel model, HumanMeshNet, that regresses a template mesh’s vertices, as well as receives a regularization by the 3D skeletal locations in a multi-branch, multi-task setup. The image to mesh vertex regression is further regularized by the neighborhood constraint imposed by mesh topology ensuring smooth surface reconstruction. The proposed paradigm can theoretically learn local surface deformations induced by body shape variations and can therefore learn high-resolution meshes going ahead. We show comparable performance with SoA (in terms of surface and joint error) with far lesser computational complexity, modeling cost and therefore real-time reconstructions on three publicly available datasets. We also show the generalizability of the proposed paradigm for a similar task of predicting hand mesh models. Given these initial results, we would like to exploit the mesh topology in an explicit manner going ahead. |
Tasks | |
Published | 2019-08-19 |
URL | https://arxiv.org/abs/1908.06544v1 |
https://arxiv.org/pdf/1908.06544v1.pdf | |
PWC | https://paperswithcode.com/paper/humanmeshnet-polygonal-mesh-recovery-of |
Repo | https://github.com/chaitanya100100/HumanMeshNet |
Framework | pytorch |
Single Point Transductive Prediction
Title | Single Point Transductive Prediction |
Authors | Nilesh Tripuraneni, Lester Mackey |
Abstract | Standard methods in supervised learning separate training and prediction: the model is fit independently of any test points it may encounter. However, can knowledge of the next test point $\mathbf{x}{\star}$ be exploited to improve prediction accuracy? We address this question in the context of linear prediction, showing how techniques from semi-parametric inference can be used transductively to combat regularization bias. We first lower bound the $\mathbf{x}{\star}$ prediction error of ridge regression and the Lasso, showing that they must incur significant bias in certain test directions. We then provide non-asymptotic upper bounds on the $\mathbf{x}_{\star}$ prediction error of two transductive prediction rules. We conclude by showing the efficacy of our methods on both synthetic and real data, highlighting the improvements single point transductive prediction can provide in settings with distribution shift. |
Tasks | |
Published | 2019-08-06 |
URL | https://arxiv.org/abs/1908.02341v3 |
https://arxiv.org/pdf/1908.02341v3.pdf | |
PWC | https://paperswithcode.com/paper/debiasing-linear-prediction |
Repo | https://github.com/nileshtrip/DebLinPredCode |
Framework | none |
Analyzing Learned Molecular Representations for Property Prediction
Title | Analyzing Learned Molecular Representations for Property Prediction |
Authors | Kevin Yang, Kyle Swanson, Wengong Jin, Connor Coley, Philipp Eiden, Hua Gao, Angel Guzman-Perez, Timothy Hopper, Brian Kelley, Miriam Mathea, Andrew Palmer, Volker Settels, Tommi Jaakkola, Klavs Jensen, Regina Barzilay |
Abstract | Advancements in neural machinery have led to a wide range of algorithmic solutions for molecular property prediction. Two classes of models in particular have yielded promising results: neural networks applied to computed molecular fingerprints or expert-crafted descriptors, and graph convolutional neural networks that construct a learned molecular representation by operating on the graph structure of the molecule. However, recent literature has yet to clearly determine which of these two methods is superior when generalizing to new chemical space. Furthermore, prior research has rarely examined these new models in industry research settings in comparison to existing employed models. In this paper, we benchmark models extensively on 19 public and 16 proprietary industrial datasets spanning a wide variety of chemical endpoints. In addition, we introduce a graph convolutional model that consistently matches or outperforms models using fixed molecular descriptors as well as previous graph neural architectures on both public and proprietary datasets. Our empirical findings indicate that while approaches based on these representations have yet to reach the level of experimental reproducibility, our proposed model nevertheless offers significant improvements over models currently used in industrial workflows. |
Tasks | Molecular Property Prediction |
Published | 2019-04-02 |
URL | https://arxiv.org/abs/1904.01561v5 |
https://arxiv.org/pdf/1904.01561v5.pdf | |
PWC | https://paperswithcode.com/paper/are-learned-molecular-representations-ready |
Repo | https://github.com/swansonk14/chemprop |
Framework | pytorch |
GMLS-Nets: A framework for learning from unstructured data
Title | GMLS-Nets: A framework for learning from unstructured data |
Authors | Nathaniel Trask, Ravi G. Patel, Ben J. Gross, Paul J. Atzberger |
Abstract | Data fields sampled on irregularly spaced points arise in many applications in the sciences and engineering. For regular grids, Convolutional Neural Networks (CNNs) have been successfully used to gaining benefits from weight sharing and invariances. We generalize CNNs by introducing methods for data on unstructured point clouds based on Generalized Moving Least Squares (GMLS). GMLS is a non-parametric technique for estimating linear bounded functionals from scattered data, and has recently been used in the literature for solving partial differential equations. By parameterizing the GMLS estimator, we obtain learning methods for operators with unstructured stencils. In GMLS-Nets the necessary calculations are local, readily parallelizable, and the estimator is supported by a rigorous approximation theory. We show how the framework may be used for unstructured physical data sets to perform functional regression to identify associated differential operators and to regress quantities of interest. The results suggest the architectures to be an attractive foundation for data-driven model development in scientific machine learning applications. |
Tasks | |
Published | 2019-09-07 |
URL | https://arxiv.org/abs/1909.05371v2 |
https://arxiv.org/pdf/1909.05371v2.pdf | |
PWC | https://paperswithcode.com/paper/gmls-nets-a-framework-for-learning-from |
Repo | https://github.com/rgp62/gmls-nets |
Framework | tf |
Adversarially Robust Distillation
Title | Adversarially Robust Distillation |
Authors | Micah Goldblum, Liam Fowl, Soheil Feizi, Tom Goldstein |
Abstract | Knowledge distillation is effective for producing small, high-performance neural networks for classification, but these small networks are vulnerable to adversarial attacks. This paper studies how adversarial robustness transfers from teacher to student during knowledge distillation. We find that a large amount of robustness may be inherited by the student even when distilled on only clean images. Second, we introduce Adversarially Robust Distillation (ARD) for distilling robustness onto student networks. In addition to producing small models with high test accuracy like conventional distillation, ARD also passes the superior robustness of large networks onto the student. In our experiments, we find that ARD student models decisively outperform adversarially trained networks of identical architecture in terms of robust accuracy, surpassing state-of-the-art methods on standard robustness benchmarks. Finally, we adapt recent fast adversarial training methods to ARD for accelerated robust distillation. |
Tasks | |
Published | 2019-05-23 |
URL | https://arxiv.org/abs/1905.09747v2 |
https://arxiv.org/pdf/1905.09747v2.pdf | |
PWC | https://paperswithcode.com/paper/190509747 |
Repo | https://github.com/goldblum/AdversariallyRobustDistillation |
Framework | pytorch |
Deep Video Inpainting
Title | Deep Video Inpainting |
Authors | Dahun Kim, Sanghyun Woo, Joon-Young Lee, In So Kweon |
Abstract | Video inpainting aims to fill spatio-temporal holes with plausible content in a video. Despite tremendous progress of deep neural networks for image inpainting, it is challenging to extend these methods to the video domain due to the additional time dimension. In this work, we propose a novel deep network architecture for fast video inpainting. Built upon an image-based encoder-decoder model, our framework is designed to collect and refine information from neighbor frames and synthesize still-unknown regions. At the same time, the output is enforced to be temporally consistent by a recurrent feedback and a temporal memory module. Compared with the state-of-the-art image inpainting algorithm, our method produces videos that are much more semantically correct and temporally smooth. In contrast to the prior video completion method which relies on time-consuming optimization, our method runs in near real-time while generating competitive video results. Finally, we applied our framework to video retargeting task, and obtain visually pleasing results. |
Tasks | Image Inpainting, Video Denoising, Video Inpainting, Video-to-Video Synthesis |
Published | 2019-05-05 |
URL | https://arxiv.org/abs/1905.01639v1 |
https://arxiv.org/pdf/1905.01639v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-video-inpainting |
Repo | https://github.com/mcahny/Deep-Video-Inpainting |
Framework | pytorch |
Twin Auxiliary Classifiers GAN
Title | Twin Auxiliary Classifiers GAN |
Authors | Mingming Gong, Yanwu Xu, Chunyuan Li, Kun Zhang, Kayhan Batmanghelich |
Abstract | Conditional generative models enjoy remarkable progress over the past few years. One of the popular conditional models is Auxiliary Classifier GAN (AC-GAN), which generates highly discriminative images by extending the loss function of GAN with an auxiliary classifier. However, the diversity of the generated samples by AC-GAN tends to decrease as the number of classes increases, hence limiting its power on large-scale data. In this paper, we identify the source of the low diversity issue theoretically and propose a practical solution to solve the problem. We show that the auxiliary classifier in AC-GAN imposes perfect separability, which is disadvantageous when the supports of the class distributions have significant overlap. To address the issue, we propose Twin Auxiliary Classifiers Generative Adversarial Net (TAC-GAN) that further benefits from a new player that interacts with other players (the generator and the discriminator) in GAN. Theoretically, we demonstrate that TAC-GAN can effectively minimize the divergence between the generated and real-data distributions. Extensive experimental results show that our TAC-GAN can successfully replicate the true data distributions on simulated data, and significantly improves the diversity of class-conditional image generation on real datasets. |
Tasks | Conditional Image Generation, Image Generation |
Published | 2019-07-05 |
URL | https://arxiv.org/abs/1907.02690v4 |
https://arxiv.org/pdf/1907.02690v4.pdf | |
PWC | https://paperswithcode.com/paper/twin-auxiliary-classifiers-gan |
Repo | https://github.com/batmanlab/twin_ac |
Framework | pytorch |
TorchGAN: A Flexible Framework for GAN Training and Evaluation
Title | TorchGAN: A Flexible Framework for GAN Training and Evaluation |
Authors | Avik Pal, Aniket Das |
Abstract | TorchGAN is a PyTorch based framework for writing succinct and comprehensible code for training and evaluation of Generative Adversarial Networks. The framework’s modular design allows effortless customization of the model architecture, loss functions, training paradigms, and evaluation metrics. The key features of TorchGAN are its extensibility, built-in support for a large number of popular models, losses and evaluation metrics, and zero overhead compared to vanilla PyTorch. By using the framework to implement several popular GAN models, we demonstrate its extensibility and ease of use. We also benchmark the training time of our framework for said models against the corresponding baseline PyTorch implementations and observe that TorchGAN’s features bear almost zero overhead. |
Tasks | Conditional Image Generation, Image Generation |
Published | 2019-09-08 |
URL | https://arxiv.org/abs/1909.03410v1 |
https://arxiv.org/pdf/1909.03410v1.pdf | |
PWC | https://paperswithcode.com/paper/torchgan-a-flexible-framework-for-gan |
Repo | https://github.com/torchgan/torchgan |
Framework | pytorch |