Paper Group AWR 197
Learning Attraction Field Representation for Robust Line Segment Detection. Malaria Detection Using Image Processing and Machine Learning. Deeply learning molecular structure-property relationships using attention- and gate-augmented graph convolutional network. DeepDTA: Deep Drug-Target Binding Affinity Prediction. High-dimension Tensor Completion …
Learning Attraction Field Representation for Robust Line Segment Detection
Title | Learning Attraction Field Representation for Robust Line Segment Detection |
Authors | Nan Xue, Song Bai, Fudong Wang, Gui-Song Xia, Tianfu Wu, Liangpei Zhang |
Abstract | This paper presents a region-partition based attraction field dual representation for line segment maps, and thus poses the problem of line segment detection (LSD) as the region coloring problem. The latter is then addressed by learning deep convolutional neural networks (ConvNets) for accuracy, robustness and efficiency. For a 2D line segment map, our dual representation consists of three components: (i) A region-partition map in which every pixel is assigned to one and only one line segment; (ii) An attraction field map in which every pixel in a partition region is encoded by its 2D projection vector w.r.t. the associated line segment; and (iii) A squeeze module which squashes the attraction field to a line segment map that almost perfectly recovers the input one. By leveraging the duality, we learn ConvNets to compute the attraction field maps for raw in-put images, followed by the squeeze module for LSD, in an end-to-end manner. Our method rigorously addresses several challenges in LSD such as local ambiguity and class imbalance. Our method also harnesses the best practices developed in ConvNets based semantic segmentation methods such as the encoder-decoder architecture and the a-trous convolution. In experiments, our method is tested on the WireFrame dataset and the YorkUrban dataset with state-of-the-art performance obtained. Especially, we advance the performance by 4.5 percents on the WireFrame dataset. Our method is also fast with 6.6~10.4 FPS, outperforming most of existing line segment detectors. |
Tasks | Line Segment Detection, Semantic Segmentation |
Published | 2018-12-05 |
URL | http://arxiv.org/abs/1812.02122v2 |
http://arxiv.org/pdf/1812.02122v2.pdf | |
PWC | https://paperswithcode.com/paper/learning-attraction-field-representation-for |
Repo | https://github.com/cherubicXN/afm_cvpr2019 |
Framework | pytorch |
Malaria Detection Using Image Processing and Machine Learning
Title | Malaria Detection Using Image Processing and Machine Learning |
Authors | Suman Kunwar |
Abstract | Malaria is mosquito-borne blood disease caused by parasites of the genus Plasmodium. Conventional diagnostic tool for malaria is the examination of stained blood cell of patient in microscope. The blood to be tested is placed in a slide and is observed under a microscope to count the number of infected RBC. An expert technician is involved in the examination of the slide with intense visual and mental concentration. This is tiresome and time consuming process. In this paper, we construct a new mage processing system for detection and quantification of plasmodium parasites in blood smear slide, later we develop Machine Learning algorithm to learn, detect and determine the types of infected cells according to its features. |
Tasks | |
Published | 2018-01-28 |
URL | http://arxiv.org/abs/1801.10031v2 |
http://arxiv.org/pdf/1801.10031v2.pdf | |
PWC | https://paperswithcode.com/paper/malaria-detection-using-image-processing-and |
Repo | https://github.com/sumn2u/react-typescript-pdf-reader |
Framework | none |
Deeply learning molecular structure-property relationships using attention- and gate-augmented graph convolutional network
Title | Deeply learning molecular structure-property relationships using attention- and gate-augmented graph convolutional network |
Authors | Seongok Ryu, Jaechang Lim, Seung Hwan Hong, Woo Youn Kim |
Abstract | Molecular structure-property relationships are key to molecular engineering for materials and drug discovery. The rise of deep learning offers a new viable solution to elucidate the structure-property relationships directly from chemical data. Here we show that the performance of graph convolutional networks (GCNs) for the prediction of molecular properties can be improved by incorporating attention and gate mechanisms. The attention mechanism enables a GCN to identify atoms in different environments. The gated skip-connection further improves the GCN by updating feature maps at an appropriate rate. We demonstrate that the resulting attention- and gate-augmented GCN could extract better structural features related to a target molecular property such as solubility, polarity, synthetic accessibility and photovoltaic efficiency compared to the vanilla GCN. More interestingly, it identified two distinct parts of molecules as essential structural features for high photovoltaic efficiency, and each of them coincided with the areas of donor and acceptor orbitals for charge-transfer excitations, respectively. As a result, the new model could accurately predict molecular properties and place molecules with similar properties close to each other in a well-trained latent space, which is critical for successful molecular engineering. |
Tasks | Drug Discovery |
Published | 2018-05-28 |
URL | http://arxiv.org/abs/1805.10988v3 |
http://arxiv.org/pdf/1805.10988v3.pdf | |
PWC | https://paperswithcode.com/paper/deeply-learning-molecular-structure-property |
Repo | https://github.com/qyuan7/molecular_gcn |
Framework | pytorch |
DeepDTA: Deep Drug-Target Binding Affinity Prediction
Title | DeepDTA: Deep Drug-Target Binding Affinity Prediction |
Authors | Hakime Öztürk, Elif Ozkirimli, Arzucan Özgür |
Abstract | The identification of novel drug-target (DT) interactions is a substantial part of the drug discovery process. Most of the computational methods that have been proposed to predict DT interactions have focused on binary classification, where the goal is to determine whether a DT pair interacts or not. However, protein-ligand interactions assume a continuum of binding strength values, also called binding affinity and predicting this value still remains a challenge. The increase in the affinity data available in DT knowledge-bases allows the use of advanced learning techniques such as deep learning architectures in the prediction of binding affinities. In this study, we propose a deep-learning based model that uses only sequence information of both targets and drugs to predict DT interaction binding affinities. The few studies that focus on DT binding affinity prediction use either 3D structures of protein-ligand complexes or 2D features of compounds. One novel approach used in this work is the modeling of protein sequences and compound 1D representations with convolutional neural networks (CNNs). The results show that the proposed deep learning based model that uses the 1D representations of targets and drugs is an effective approach for drug target binding affinity prediction. The model in which high-level representations of a drug and a target are constructed via CNNs achieved the best Concordance Index (CI) performance in one of our larger benchmark data sets, outperforming the KronRLS algorithm and SimBoost, a state-of-the-art method for DT binding affinity prediction. |
Tasks | Drug Discovery |
Published | 2018-01-30 |
URL | http://arxiv.org/abs/1801.10193v2 |
http://arxiv.org/pdf/1801.10193v2.pdf | |
PWC | https://paperswithcode.com/paper/deepdta-deep-drug-target-binding-affinity |
Repo | https://github.com/Yindong-Zhang/GraphConvolutionDrugTargetInteration |
Framework | tf |
High-dimension Tensor Completion via Gradient-based Optimization Under Tensor-train Format
Title | High-dimension Tensor Completion via Gradient-based Optimization Under Tensor-train Format |
Authors | Longhao Yuan, Qibin Zhao, Lihua Gui, Jianting Cao |
Abstract | Tensor train (TT) decomposition has drawn people’s attention due to its powerful representation ability and performance stability in high-order tensors. In this paper, we propose a novel approach to recover the missing entries of incomplete data represented by higher-order tensors. We attempt to find the low-rank TT decomposition of the incomplete data which captures the latent features of the whole data and then reconstruct the missing entries. By applying gradient descent algorithms, tensor completion problem is efficiently solved by optimization models. We propose two TT-based algorithms: Tensor Train Weighted Optimization (TT-WOPT) and Tensor Train Stochastic Gradient Descent (TT-SGD) to optimize TT decomposition factors. In addition, a method named Visual Data Tensorization (VDT) is proposed to transform visual data into higher-order tensors, resulting in the performance improvement of our algorithms. The experiments in synthetic data and visual data show high efficiency and performance of our algorithms compared to the state-of-the-art completion algorithms, especially in high-order, high missing rate, and large-scale tensor completion situations. |
Tasks | |
Published | 2018-04-05 |
URL | http://arxiv.org/abs/1804.01983v3 |
http://arxiv.org/pdf/1804.01983v3.pdf | |
PWC | https://paperswithcode.com/paper/high-dimension-tensor-completion-via-gradient |
Repo | https://github.com/yuanlonghao/T3C_tensor_completion |
Framework | none |
Tensor Ring Decomposition with Rank Minimization on Latent Space: An Efficient Approach for Tensor Completion
Title | Tensor Ring Decomposition with Rank Minimization on Latent Space: An Efficient Approach for Tensor Completion |
Authors | Longhao Yuan, Chao Li, Danilo Mandic, Jianting Cao, Qibin Zhao |
Abstract | In tensor completion tasks, the traditional low-rank tensor decomposition models suffer from the laborious model selection problem due to their high model sensitivity. In particular, for tensor ring (TR) decomposition, the number of model possibilities grows exponentially with the tensor order, which makes it rather challenging to find the optimal TR decomposition. In this paper, by exploiting the low-rank structure of the TR latent space, we propose a novel tensor completion method which is robust to model selection. In contrast to imposing the low-rank constraint on the data space, we introduce nuclear norm regularization on the latent TR factors, resulting in the optimization step using singular value decomposition (SVD) being performed at a much smaller scale. By leveraging the alternating direction method of multipliers (ADMM) scheme, the latent TR factors with optimal rank and the recovered tensor can be obtained simultaneously. Our proposed algorithm is shown to effectively alleviate the burden of TR-rank selection, thereby greatly reducing the computational cost. The extensive experimental results on both synthetic and real-world data demonstrate the superior performance and efficiency of the proposed approach against the state-of-the-art algorithms. |
Tasks | Model Selection |
Published | 2018-09-07 |
URL | http://arxiv.org/abs/1809.02288v2 |
http://arxiv.org/pdf/1809.02288v2.pdf | |
PWC | https://paperswithcode.com/paper/tensor-ring-decomposition-with-rank |
Repo | https://github.com/zhaoxile/reproducible-tensor-completion-state-of-the-art |
Framework | none |
Representation Flow for Action Recognition
Title | Representation Flow for Action Recognition |
Authors | AJ Piergiovanni, Michael S. Ryoo |
Abstract | In this paper, we propose a convolutional layer inspired by optical flow algorithms to learn motion representations. Our representation flow layer is a fully-differentiable layer designed to capture the flow' of any representation channel within a convolutional neural network for action recognition. Its parameters for iterative flow optimization are learned in an end-to-end fashion together with the other CNN model parameters, maximizing the action recognition performance. Furthermore, we newly introduce the concept of learning flow of flow’ representations by stacking multiple representation flow layers. We conducted extensive experimental evaluations, confirming its advantages over previous recognition models using traditional optical flows in both computational speed and performance. Code/models available here: https://piergiaj.github.io/rep-flow-site/ |
Tasks | Action Classification, Action Recognition In Videos, Activity Recognition, Activity Recognition In Videos, Optical Flow Estimation, Temporal Action Localization, Video Classification, Video Understanding |
Published | 2018-10-02 |
URL | https://arxiv.org/abs/1810.01455v3 |
https://arxiv.org/pdf/1810.01455v3.pdf | |
PWC | https://paperswithcode.com/paper/representation-flow-for-action-recognition |
Repo | https://github.com/piergiaj/representation-flow-cvpr19 |
Framework | pytorch |
Hybrid Binary Networks: Optimizing for Accuracy, Efficiency and Memory
Title | Hybrid Binary Networks: Optimizing for Accuracy, Efficiency and Memory |
Authors | Ameya Prabhu, Vishal Batchu, Rohit Gajawada, Sri Aurobindo Munagala, Anoop Namboodiri |
Abstract | Binarization is an extreme network compression approach that provides large computational speedups along with energy and memory savings, albeit at significant accuracy costs. We investigate the question of where to binarize inputs at layer-level granularity and show that selectively binarizing the inputs to specific layers in the network could lead to significant improvements in accuracy while preserving most of the advantages of binarization. We analyze the binarization tradeoff using a metric that jointly models the input binarization-error and computational cost and introduce an efficient algorithm to select layers whose inputs are to be binarized. Practical guidelines based on insights obtained from applying the algorithm to a variety of models are discussed. Experiments on Imagenet dataset using AlexNet and ResNet-18 models show 3-4% improvements in accuracy over fully binarized networks with minimal impact on compression and computational speed. The improvements are even more substantial on sketch datasets like TU-Berlin, where we match state-of-the-art accuracy as well, getting over 8% increase in accuracies. We further show that our approach can be applied in tandem with other forms of compression that deal with individual layers or overall model compression (e.g., SqueezeNets). Unlike previous quantization approaches, we are able to binarize the weights in the last layers of a network, which often have a large number of parameters, resulting in significant improvement in accuracy over fully binarized models. |
Tasks | Model Compression, Quantization |
Published | 2018-04-11 |
URL | http://arxiv.org/abs/1804.03867v1 |
http://arxiv.org/pdf/1804.03867v1.pdf | |
PWC | https://paperswithcode.com/paper/hybrid-binary-networks-optimizing-for |
Repo | https://github.com/erilyth/HybridBinaryNetworks-WACV18 |
Framework | pytorch |
Invariant Representations without Adversarial Training
Title | Invariant Representations without Adversarial Training |
Authors | Daniel Moyer, Shuyang Gao, Rob Brekelmans, Greg Ver Steeg, Aram Galstyan |
Abstract | Representations of data that are invariant to changes in specified factors are useful for a wide range of problems: removing potential biases in prediction problems, controlling the effects of covariates, and disentangling meaningful factors of variation. Unfortunately, learning representations that exhibit invariance to arbitrary nuisance factors yet remain useful for other tasks is challenging. Existing approaches cast the trade-off between task performance and invariance in an adversarial way, using an iterative minimax optimization. We show that adversarial training is unnecessary and sometimes counter-productive; we instead cast invariant representation learning as a single information-theoretic objective that can be directly optimized. We demonstrate that this approach matches or exceeds performance of state-of-the-art adversarial approaches for learning fair representations and for generative modeling with controllable transformations. |
Tasks | Representation Learning |
Published | 2018-05-24 |
URL | https://arxiv.org/abs/1805.09458v4 |
https://arxiv.org/pdf/1805.09458v4.pdf | |
PWC | https://paperswithcode.com/paper/invariant-representations-without-adversarial |
Repo | https://github.com/dcmoyer/inv-rep |
Framework | tf |
Hierarchical Deep Multiagent Reinforcement Learning with Temporal Abstraction
Title | Hierarchical Deep Multiagent Reinforcement Learning with Temporal Abstraction |
Authors | Hongyao Tang, Jianye Hao, Tangjie Lv, Yingfeng Chen, Zongzhang Zhang, Hangtian Jia, Chunxu Ren, Yan Zheng, Zhaopeng Meng, Changjie Fan, Li Wang |
Abstract | Multiagent reinforcement learning (MARL) is commonly considered to suffer from non-stationary environments and exponentially increasing policy space. It would be even more challenging when rewards are sparse and delayed over long trajectories. In this paper, we study hierarchical deep MARL in cooperative multiagent problems with sparse and delayed reward. With temporal abstraction, we decompose the problem into a hierarchy of different time scales and investigate how agents can learn high-level coordination based on the independent skills learned at the low level. Three hierarchical deep MARL architectures are proposed to learn hierarchical policies under different MARL paradigms. Besides, we propose a new experience replay mechanism to alleviate the issue of the sparse transitions at the high level of abstraction and the non-stationarity of multiagent learning. We empirically demonstrate the effectiveness of our approaches in two domains with extremely sparse feedback: (1) a variety of Multiagent Trash Collection tasks, and (2) a challenging online mobile game, i.e., Fever Basketball Defense. |
Tasks | |
Published | 2018-09-25 |
URL | https://arxiv.org/abs/1809.09332v2 |
https://arxiv.org/pdf/1809.09332v2.pdf | |
PWC | https://paperswithcode.com/paper/hierarchical-deep-multiagent-reinforcement |
Repo | https://github.com/bluecontra/MATC_Env |
Framework | none |
Decorrelated Batch Normalization
Title | Decorrelated Batch Normalization |
Authors | Lei Huang, Dawei Yang, Bo Lang, Jia Deng |
Abstract | Batch Normalization (BN) is capable of accelerating the training of deep models by centering and scaling activations within mini-batches. In this work, we propose Decorrelated Batch Normalization (DBN), which not just centers and scales activations but whitens them. We explore multiple whitening techniques, and find that PCA whitening causes a problem we call stochastic axis swapping, which is detrimental to learning. We show that ZCA whitening does not suffer from this problem, permitting successful learning. DBN retains the desirable qualities of BN and further improves BN’s optimization efficiency and generalization ability. We design comprehensive experiments to show that DBN can improve the performance of BN on multilayer perceptrons and convolutional neural networks. Furthermore, we consistently improve the accuracy of residual networks on CIFAR-10, CIFAR-100, and ImageNet. |
Tasks | |
Published | 2018-04-23 |
URL | http://arxiv.org/abs/1804.08450v1 |
http://arxiv.org/pdf/1804.08450v1.pdf | |
PWC | https://paperswithcode.com/paper/decorrelated-batch-normalization |
Repo | https://github.com/bhneo/DecorrelatedBN_tf |
Framework | tf |
CariGANs: Unpaired Photo-to-Caricature Translation
Title | CariGANs: Unpaired Photo-to-Caricature Translation |
Authors | Kaidi Cao, Jing Liao, Lu Yuan |
Abstract | Facial caricature is an art form of drawing faces in an exaggerated way to convey humor or sarcasm. In this paper, we propose the first Generative Adversarial Network (GAN) for unpaired photo-to-caricature translation, which we call “CariGANs”. It explicitly models geometric exaggeration and appearance stylization using two components: CariGeoGAN, which only models the geometry-to-geometry transformation from face photos to caricatures, and CariStyGAN, which transfers the style appearance from caricatures to face photos without any geometry deformation. In this way, a difficult cross-domain translation problem is decoupled into two easier tasks. The perceptual study shows that caricatures generated by our CariGANs are closer to the hand-drawn ones, and at the same time better persevere the identity, compared to state-of-the-art methods. Moreover, our CariGANs allow users to control the shape exaggeration degree and change the color/texture style by tuning the parameters or giving an example caricature. |
Tasks | Caricature, Photo-To-Caricature Translation |
Published | 2018-11-01 |
URL | http://arxiv.org/abs/1811.00222v2 |
http://arxiv.org/pdf/1811.00222v2.pdf | |
PWC | https://paperswithcode.com/paper/carigans-unpaired-photo-to-caricature |
Repo | https://github.com/SerialLain3170/ComputerVision-Papers |
Framework | none |
SQL-to-Text Generation with Graph-to-Sequence Model
Title | SQL-to-Text Generation with Graph-to-Sequence Model |
Authors | Kun Xu, Lingfei Wu, Zhiguo Wang, Yansong Feng, Vadim Sheinin |
Abstract | Previous work approaches the SQL-to-text generation task using vanilla Seq2Seq models, which may not fully capture the inherent graph-structured information in SQL query. In this paper, we first introduce a strategy to represent the SQL query as a directed graph and then employ a graph-to-sequence model to encode the global structure information into node embeddings. This model can effectively learn the correlation between the SQL query pattern and its interpretation. Experimental results on the WikiSQL dataset and Stackoverflow dataset show that our model significantly outperforms the Seq2Seq and Tree2Seq baselines, achieving the state-of-the-art performance. |
Tasks | Graph-to-Sequence, SQL-to-Text, Text Generation |
Published | 2018-09-14 |
URL | http://arxiv.org/abs/1809.05255v2 |
http://arxiv.org/pdf/1809.05255v2.pdf | |
PWC | https://paperswithcode.com/paper/sql-to-text-generation-with-graph-to-sequence |
Repo | https://github.com/IBM/SQL-to-Text |
Framework | tf |
Arc-support Line Segments Revisited: An Efficient and High-quality Ellipse Detection
Title | Arc-support Line Segments Revisited: An Efficient and High-quality Ellipse Detection |
Authors | Changsheng Lu, Siyu Xia, Ming Shao, Yun Fu |
Abstract | Over the years many ellipse detection algorithms spring up and are studied broadly, while the critical issue of detecting ellipses accurately and efficiently in real-world images remains a challenge. In this paper, we propose a valuable industry-oriented ellipse detector by arc-support line segments, which simultaneously reaches high detection accuracy and efficiency. To simplify the complicated curves in an image while retaining the general properties including convexity and polarity, the arc-support line segments are extracted, which grounds the successful detection of ellipses. The arc-support groups are formed by iteratively and robustly linking the arc-support line segments that latently belong to a common ellipse. Afterward, two complementary approaches, namely, locally selecting the arc-support group with higher saliency and globally searching all the valid paired groups, are adopted to fit the initial ellipses in a fast way. Then, the ellipse candidate set can be formulated by hierarchical clustering of 5D parameter space of initial ellipses. Finally, the salient ellipse candidates are selected and refined as detections subject to the stringent and effective verification. Extensive experiments on three public datasets are implemented and our method achieves the best F-measure scores compared to the state-of-the-art methods. The source code is available at https://github.com/AlanLuSun/High-quality-ellipse-detection. |
Tasks | |
Published | 2018-10-08 |
URL | https://arxiv.org/abs/1810.03243v5 |
https://arxiv.org/pdf/1810.03243v5.pdf | |
PWC | https://paperswithcode.com/paper/arc-support-line-segments-revisited-an |
Repo | https://github.com/AlanLuSun/High-quality-ellipse-detection |
Framework | none |
ServeNet: A Deep Neural Network for Web Service Classification
Title | ServeNet: A Deep Neural Network for Web Service Classification |
Authors | Yilong Yang, Wei Ke, Peng Liu, Weiru Wang, Bingqing Shen, Bo Liu, Yongxin Zhao |
Abstract | Automated service classification plays a crucial role in service discovery, selection, and composition. Machine learning has been used for service classification in recent years. However, the performance of conventional machine learning methods highly depends on the quality of manual feature engineering. In this paper, we present a deep neural network to automatically abstract low-level representation of service description to high-level features without feature engineering and then predict service classification on 50 service categories. To demonstrate the effectiveness of our approach, we conduct a comprehensive experimental study by comparing 10 machine learning methods on 10,000 real-world web services. The result shows that the proposed deep neural network can achieve higher accuracy than other machine learning methods. |
Tasks | Feature Engineering |
Published | 2018-06-14 |
URL | https://arxiv.org/abs/1806.05437v2 |
https://arxiv.org/pdf/1806.05437v2.pdf | |
PWC | https://paperswithcode.com/paper/servenet-a-deep-neural-network-for-web |
Repo | https://github.com/yylonly/ServeNet |
Framework | none |