Paper Group AWR 185
Classifying Online Dating Profiles on Tinder using FaceNet Facial Embeddings. Defining Locality for Surrogates in Post-hoc Interpretablity. Baseline Needs More Love: On Simple Word-Embedding-Based Models and Associated Pooling Mechanisms. DeepLiDAR: Deep Surface Normal Guided Depth Prediction for Outdoor Scene from Sparse LiDAR Data and Single Colo …
Classifying Online Dating Profiles on Tinder using FaceNet Facial Embeddings
Title | Classifying Online Dating Profiles on Tinder using FaceNet Facial Embeddings |
Authors | Charles F Jekel, Raphael T. Haftka |
Abstract | A method to produce personalized classification models to automatically review online dating profiles on Tinder is proposed, based on the user’s historical preference. The method takes advantage of a FaceNet facial classification model to extract features which may be related to facial attractiveness. The embeddings from a FaceNet model were used as the features to describe an individual’s face. A user reviewed 8,545 online dating profiles. For each reviewed online dating profile, a feature set was constructed from the profile images which contained just one face. Two approaches are presented to go from the set of features for each face, to a set of profile features. A simple logistic regression trained on the embeddings from just 20 profiles could obtain a 65% validation accuracy. A point of diminishing marginal returns was identified to occur around 80 profiles, at which the model accuracy of 73% would only improve marginally after reviewing a significant number of additional profiles. |
Tasks | |
Published | 2018-03-12 |
URL | http://arxiv.org/abs/1803.04347v1 |
http://arxiv.org/pdf/1803.04347v1.pdf | |
PWC | https://paperswithcode.com/paper/classifying-online-dating-profiles-on-tinder |
Repo | https://github.com/cjekel/tindetheus |
Framework | tf |
Defining Locality for Surrogates in Post-hoc Interpretablity
Title | Defining Locality for Surrogates in Post-hoc Interpretablity |
Authors | Thibault Laugel, Xavier Renard, Marie-Jeanne Lesot, Christophe Marsala, Marcin Detyniecki |
Abstract | Local surrogate models, to approximate the local decision boundary of a black-box classifier, constitute one approach to generate explanations for the rationale behind an individual prediction made by the back-box. This paper highlights the importance of defining the right locality, the neighborhood on which a local surrogate is trained, in order to approximate accurately the local black-box decision boundary. Unfortunately, as shown in this paper, this issue is not only a parameter or sampling distribution challenge and has a major impact on the relevance and quality of the approximation of the local black-box decision boundary and thus on the meaning and accuracy of the generated explanation. To overcome the identified problems, quantified with an adapted measure and procedure, we propose to generate surrogate-based explanations for individual predictions based on a sampling centered on particular place of the decision boundary, relevant for the prediction to be explained, rather than on the prediction itself as it is classically done. We evaluate the novel approach compared to state-of-the-art methods and a straightforward improvement thereof on four UCI datasets. |
Tasks | |
Published | 2018-06-19 |
URL | http://arxiv.org/abs/1806.07498v1 |
http://arxiv.org/pdf/1806.07498v1.pdf | |
PWC | https://paperswithcode.com/paper/defining-locality-for-surrogates-in-post-hoc |
Repo | https://github.com/axadil/locality-interpretable-surrogate |
Framework | none |
Baseline Needs More Love: On Simple Word-Embedding-Based Models and Associated Pooling Mechanisms
Title | Baseline Needs More Love: On Simple Word-Embedding-Based Models and Associated Pooling Mechanisms |
Authors | Dinghan Shen, Guoyin Wang, Wenlin Wang, Martin Renqiang Min, Qinliang Su, Yizhe Zhang, Chunyuan Li, Ricardo Henao, Lawrence Carin |
Abstract | Many deep learning architectures have been proposed to model the compositionality in text sequences, requiring a substantial number of parameters and expensive computations. However, there has not been a rigorous evaluation regarding the added value of sophisticated compositional functions. In this paper, we conduct a point-by-point comparative study between Simple Word-Embedding-based Models (SWEMs), consisting of parameter-free pooling operations, relative to word-embedding-based RNN/CNN models. Surprisingly, SWEMs exhibit comparable or even superior performance in the majority of cases considered. Based upon this understanding, we propose two additional pooling strategies over learned word embeddings: (i) a max-pooling operation for improved interpretability; and (ii) a hierarchical pooling operation, which preserves spatial (n-gram) information within text sequences. We present experiments on 17 datasets encompassing three tasks: (i) (long) document classification; (ii) text sequence matching; and (iii) short text tasks, including classification and tagging. The source code and datasets can be obtained from https:// github.com/dinghanshen/SWEM. |
Tasks | Document Classification, Named Entity Recognition, Sentiment Analysis, Subjectivity Analysis, Text Classification, Word Embeddings |
Published | 2018-05-24 |
URL | http://arxiv.org/abs/1805.09843v1 |
http://arxiv.org/pdf/1805.09843v1.pdf | |
PWC | https://paperswithcode.com/paper/baseline-needs-more-love-on-simple-word |
Repo | https://github.com/nyk510/scdv-python |
Framework | none |
DeepLiDAR: Deep Surface Normal Guided Depth Prediction for Outdoor Scene from Sparse LiDAR Data and Single Color Image
Title | DeepLiDAR: Deep Surface Normal Guided Depth Prediction for Outdoor Scene from Sparse LiDAR Data and Single Color Image |
Authors | Jiaxiong Qiu, Zhaopeng Cui, Yinda Zhang, Xingdi Zhang, Shuaicheng Liu, Bing Zeng, Marc Pollefeys |
Abstract | In this paper, we propose a deep learning architecture that produces accurate dense depth for the outdoor scene from a single color image and a sparse depth. Inspired by the indoor depth completion, our network estimates surface normals as the intermediate representation to produce dense depth, and can be trained end-to-end. With a modified encoder-decoder structure, our network effectively fuses the dense color image and the sparse LiDAR depth. To address outdoor specific challenges, our network predicts a confidence mask to handle mixed LiDAR signals near foreground boundaries due to occlusion, and combines estimates from the color image and surface normals with learned attention maps to improve the depth accuracy especially for distant areas. Extensive experiments demonstrate that our model improves upon the state-of-the-art performance on KITTI depth completion benchmark. Ablation study shows the positive impact of each model components to the final performance, and comprehensive analysis shows that our model generalizes well to the input with higher sparsity or from indoor scenes. |
Tasks | Depth Completion, Depth Estimation |
Published | 2018-12-02 |
URL | http://arxiv.org/abs/1812.00488v2 |
http://arxiv.org/pdf/1812.00488v2.pdf | |
PWC | https://paperswithcode.com/paper/deeplidar-deep-surface-normal-guided-depth |
Repo | https://github.com/JiaxiongQ/DeepLiDAR |
Framework | pytorch |
Adversarial Reprogramming of Text Classification Neural Networks
Title | Adversarial Reprogramming of Text Classification Neural Networks |
Authors | Paarth Neekhara, Shehzeen Hussain, Shlomo Dubnov, Farinaz Koushanfar |
Abstract | Adversarial Reprogramming has demonstrated success in utilizing pre-trained neural network classifiers for alternative classification tasks without modification to the original network. An adversary in such an attack scenario trains an additive contribution to the inputs to repurpose the neural network for the new classification task. While this reprogramming approach works for neural networks with a continuous input space such as that of images, it is not directly applicable to neural networks trained for tasks such as text classification, where the input space is discrete. Repurposing such classification networks would require the attacker to learn an adversarial program that maps inputs from one discrete space to the other. In this work, we introduce a context-based vocabulary remapping model to reprogram neural networks trained on a specific sequence classification task, for a new sequence classification task desired by the adversary. We propose training procedures for this adversarial program in both white-box and black-box settings. We demonstrate the application of our model by adversarially repurposing various text-classification models including LSTM, bi-directional LSTM and CNN for alternate classification tasks. |
Tasks | Text Classification |
Published | 2018-09-06 |
URL | https://arxiv.org/abs/1809.01829v4 |
https://arxiv.org/pdf/1809.01829v4.pdf | |
PWC | https://paperswithcode.com/paper/adversarial-reprogramming-of-sequence |
Repo | https://github.com/paarthneekhara/rnn_adversarial_reprogramming |
Framework | pytorch |
Updating the generator in PPGN-h with gradients flowing through the encoder
Title | Updating the generator in PPGN-h with gradients flowing through the encoder |
Authors | Hesam Pakdaman |
Abstract | The Generative Adversarial Network framework has shown success in implicitly modeling data distributions and is able to generate realistic samples. Its architecture is comprised of a generator, which produces fake data that superficially seem to belong to the real data distribution, and a discriminator which is to distinguish fake from genuine samples. The Noiseless Joint Plug & Play model offers an extension to the framework by simultaneously training autoencoders. This model uses a pre-trained encoder as a feature extractor, feeding the generator with global information. Using the Plug & Play network as baseline, we design a new model by adding discriminators to the Plug & Play architecture. These additional discriminators are trained to discern real and fake latent codes, which are the output of the encoder using genuine and generated inputs, respectively. We proceed to investigate whether this approach is viable. Experiments conducted for the MNIST manifold show that this indeed is the case. |
Tasks | |
Published | 2018-04-02 |
URL | http://arxiv.org/abs/1804.00630v1 |
http://arxiv.org/pdf/1804.00630v1.pdf | |
PWC | https://paperswithcode.com/paper/updating-the-generator-in-ppgn-h-with |
Repo | https://github.com/hesampakdaman/ppgn-disc |
Framework | pytorch |
Scaling Submodular Optimization Approaches for Control Applications in Networked Systems
Title | Scaling Submodular Optimization Approaches for Control Applications in Networked Systems |
Authors | Arun V Sathanur |
Abstract | Often times, in many design problems, there is a need to select a small set of informative or representative elements from a large ground set of entities in an optimal fashion. Submodular optimization that provides for a formal way to solve such problems, has recently received significant attention from the controls community where such subset selection problems are abound. However, scaling these approaches to large systems can be challenging because of the high computational complexity of the overall flow, in-part due to the high-complexity compute-oracles used to determine the objective function values. In this work, we explore a well-known paradigm, namely leader-selection in a multi-agent networked environment to illustrate strategies for scalable submodular optimization. We study the performance of the state-of-the-art stochastic and distributed greedy algorithms as well as explore techniques that accelerate the computation oracles within the optimization loop. We finally present results combining accelerated greedy algorithms with accelerated computation oracles and demonstrate significant speedups with little loss of optimality when compared to the baseline ordinary greedy algorithm. |
Tasks | |
Published | 2018-10-05 |
URL | http://arxiv.org/abs/1810.02837v1 |
http://arxiv.org/pdf/1810.02837v1.pdf | |
PWC | https://paperswithcode.com/paper/scaling-submodular-optimization-approaches |
Repo | https://github.com/arunsv/submodular |
Framework | none |
Taking Human out of Learning Applications: A Survey on Automated Machine Learning
Title | Taking Human out of Learning Applications: A Survey on Automated Machine Learning |
Authors | Quanming Yao, Mengshuo Wang, Yuqiang Chen, Wenyuan Dai, Yu-Feng Li, Wei-Wei Tu, Qiang Yang, Yang Yu |
Abstract | Machine learning techniques have deeply rooted in our everyday life. However, since it is knowledge- and labor-intensive to pursue good learning performance, human experts are heavily involved in every aspect of machine learning. In order to make machine learning techniques easier to apply and reduce the demand for experienced human experts, automated machine learning (AutoML) has emerged as a hot topic with both industrial and academic interest. In this paper, we provide an up to date survey on AutoML. First, we introduce and define the AutoML problem, with inspiration from both realms of automation and machine learning. Then, we propose a general AutoML framework that not only covers most existing approaches to date but also can guide the design for new methods. Subsequently, we categorize and review the existing works from two aspects, i.e., the problem setup and the employed techniques. Finally, we provide a detailed analysis of AutoML approaches and explain the reasons underneath their successful applications. We hope this survey can serve as not only an insightful guideline for AutoML beginners but also an inspiration for future research. |
Tasks | AutoML |
Published | 2018-10-31 |
URL | https://arxiv.org/abs/1810.13306v4 |
https://arxiv.org/pdf/1810.13306v4.pdf | |
PWC | https://paperswithcode.com/paper/taking-human-out-of-learning-applications-a |
Repo | https://github.com/xiangning-chen/SIF |
Framework | pytorch |
An Empirical Study of Rich Subgroup Fairness for Machine Learning
Title | An Empirical Study of Rich Subgroup Fairness for Machine Learning |
Authors | Michael Kearns, Seth Neel, Aaron Roth, Zhiwei Steven Wu |
Abstract | Kearns et al. [2018] recently proposed a notion of rich subgroup fairness intended to bridge the gap between statistical and individual notions of fairness. Rich subgroup fairness picks a statistical fairness constraint (say, equalizing false positive rates across protected groups), but then asks that this constraint hold over an exponentially or infinitely large collection of subgroups defined by a class of functions with bounded VC dimension. They give an algorithm guaranteed to learn subject to this constraint, under the condition that it has access to oracles for perfectly learning absent a fairness constraint. In this paper, we undertake an extensive empirical evaluation of the algorithm of Kearns et al. On four real datasets for which fairness is a concern, we investigate the basic convergence of the algorithm when instantiated with fast heuristics in place of learning oracles, measure the tradeoffs between fairness and accuracy, and compare this approach with the recent algorithm of Agarwal et al. [2018], which implements weaker and more traditional marginal fairness constraints defined by individual protected attributes. We find that in general, the Kearns et al. algorithm converges quickly, large gains in fairness can be obtained with mild costs to accuracy, and that optimizing accuracy subject only to marginal fairness leads to classifiers with substantial subgroup unfairness. We also provide a number of analyses and visualizations of the dynamics and behavior of the Kearns et al. algorithm. Overall we find this algorithm to be effective on real data, and rich subgroup fairness to be a viable notion in practice. |
Tasks | |
Published | 2018-08-24 |
URL | http://arxiv.org/abs/1808.08166v1 |
http://arxiv.org/pdf/1808.08166v1.pdf | |
PWC | https://paperswithcode.com/paper/an-empirical-study-of-rich-subgroup-fairness |
Repo | https://github.com/SaeedSharifiMa/FairDP |
Framework | none |
Content-Based Citation Recommendation
Title | Content-Based Citation Recommendation |
Authors | Chandra Bhagavatula, Sergey Feldman, Russell Power, Waleed Ammar |
Abstract | We present a content-based method for recommending citations in an academic paper draft. We embed a given query document into a vector space, then use its nearest neighbors as candidates, and rerank the candidates using a discriminative model trained to distinguish between observed and unobserved citations. Unlike previous work, our method does not require metadata such as author names which can be missing, e.g., during the peer review process. Without using metadata, our method outperforms the best reported results on PubMed and DBLP datasets with relative improvements of over 18% in F1@20 and over 22% in MRR. We show empirically that, although adding metadata improves the performance on standard metrics, it favors self-citations which are less useful in a citation recommendation setup. We release an online portal (http://labs.semanticscholar.org/citeomatic/) for citation recommendation based on our method, and a new dataset OpenCorpus of 7 million research articles to facilitate future research on this task. |
Tasks | |
Published | 2018-02-22 |
URL | http://arxiv.org/abs/1802.08301v1 |
http://arxiv.org/pdf/1802.08301v1.pdf | |
PWC | https://paperswithcode.com/paper/content-based-citation-recommendation |
Repo | https://github.com/allenai/citeomatic |
Framework | tf |
Context-aware Deep Feature Compression for High-speed Visual Tracking
Title | Context-aware Deep Feature Compression for High-speed Visual Tracking |
Authors | Jongwon Choi, Hyung Jin Chang, Tobias Fischer, Sangdoo Yun, Kyuewang Lee, Jiyeoup Jeong, Yiannis Demiris, Jin Young Choi |
Abstract | We propose a new context-aware correlation filter based tracking framework to achieve both high computational speed and state-of-the-art performance among real-time trackers. The major contribution to the high computational speed lies in the proposed deep feature compression that is achieved by a context-aware scheme utilizing multiple expert auto-encoders; a context in our framework refers to the coarse category of the tracking target according to appearance patterns. In the pre-training phase, one expert auto-encoder is trained per category. In the tracking phase, the best expert auto-encoder is selected for a given target, and only this auto-encoder is used. To achieve high tracking performance with the compressed feature map, we introduce extrinsic denoising processes and a new orthogonality loss term for pre-training and fine-tuning of the expert auto-encoders. We validate the proposed context-aware framework through a number of experiments, where our method achieves a comparable performance to state-of-the-art trackers which cannot run in real-time, while running at a significantly fast speed of over 100 fps. |
Tasks | Denoising, Visual Object Tracking, Visual Tracking |
Published | 2018-03-28 |
URL | http://arxiv.org/abs/1803.10537v1 |
http://arxiv.org/pdf/1803.10537v1.pdf | |
PWC | https://paperswithcode.com/paper/context-aware-deep-feature-compression-for |
Repo | https://github.com/jongwon20000/TRACA |
Framework | none |
Olive Oil is Made of Olives, Baby Oil is Made for Babies: Interpreting Noun Compounds using Paraphrases in a Neural Model
Title | Olive Oil is Made of Olives, Baby Oil is Made for Babies: Interpreting Noun Compounds using Paraphrases in a Neural Model |
Authors | Vered Shwartz, Chris Waterson |
Abstract | Automatic interpretation of the relation between the constituents of a noun compound, e.g. olive oil (source) and baby oil (purpose) is an important task for many NLP applications. Recent approaches are typically based on either noun-compound representations or paraphrases. While the former has initially shown promising results, recent work suggests that the success stems from memorizing single prototypical words for each relation. We explore a neural paraphrasing approach that demonstrates superior performance when such memorization is not possible. |
Tasks | |
Published | 2018-03-21 |
URL | http://arxiv.org/abs/1803.08073v1 |
http://arxiv.org/pdf/1803.08073v1.pdf | |
PWC | https://paperswithcode.com/paper/olive-oil-is-made-of-olives-baby-oil-is-made |
Repo | https://github.com/tensorflow/models/tree/master/research/lexnet_nc |
Framework | tf |
Deep Segmentation and Registration in X-Ray Angiography Video
Title | Deep Segmentation and Registration in X-Ray Angiography Video |
Authors | Athanasios Vlontzos, Krystian Mikolajczyk |
Abstract | In interventional radiology, short video sequences of vein structure in motion are captured in order to help medical personnel identify vascular issues or plan intervention. Semantic segmentation can greatly improve the usefulness of these videos by indicating exact position of vessels and instruments, thus reducing the ambiguity. We propose a real-time segmentation method for these tasks, based on U-Net network trained in a Siamese architecture from automatically generated annotations. We make use of noisy low level binary segmentation and optical flow to generate multi class annotations that are successively improved in a multistage segmentation approach. We significantly improve the performance of a state of the art U-Net at the processing speeds of 90fps. |
Tasks | Optical Flow Estimation, Semantic Segmentation |
Published | 2018-05-16 |
URL | http://arxiv.org/abs/1805.06406v2 |
http://arxiv.org/pdf/1805.06406v2.pdf | |
PWC | https://paperswithcode.com/paper/deep-segmentation-and-registration-in-x-ray |
Repo | https://github.com/thanosvlo/Deep-Segmentation-and-Registration-in-X-Ray-Angiography-Video |
Framework | none |
GuacaMol: Benchmarking Models for De Novo Molecular Design
Title | GuacaMol: Benchmarking Models for De Novo Molecular Design |
Authors | Nathan Brown, Marco Fiscato, Marwin H. S. Segler, Alain C. Vaucher |
Abstract | De novo design seeks to generate molecules with required property profiles by virtual design-make-test cycles. With the emergence of deep learning and neural generative models in many application areas, models for molecular design based on neural networks appeared recently and show promising results. However, the new models have not been profiled on consistent tasks, and comparative studies to well-established algorithms have only seldom been performed. To standardize the assessment of both classical and neural models for de novo molecular design, we propose an evaluation framework, GuacaMol, based on a suite of standardized benchmarks. The benchmark tasks encompass measuring the fidelity of the models to reproduce the property distribution of the training sets, the ability to generate novel molecules, the exploration and exploitation of chemical space, and a variety of single and multi-objective optimization tasks. The benchmarking open-source Python code, and a leaderboard can be found on https://benevolent.ai/guacamol |
Tasks | Drug Discovery |
Published | 2018-11-22 |
URL | http://arxiv.org/abs/1811.09621v2 |
http://arxiv.org/pdf/1811.09621v2.pdf | |
PWC | https://paperswithcode.com/paper/guacamol-benchmarking-models-for-de-novo |
Repo | https://github.com/benevolentAI/guacamol_baselines |
Framework | pytorch |
ABC: A Big CAD Model Dataset For Geometric Deep Learning
Title | ABC: A Big CAD Model Dataset For Geometric Deep Learning |
Authors | Sebastian Koch, Albert Matveev, Zhongshi Jiang, Francis Williams, Alexey Artemov, Evgeny Burnaev, Marc Alexa, Denis Zorin, Daniele Panozzo |
Abstract | We introduce ABC-Dataset, a collection of one million Computer-Aided Design (CAD) models for research of geometric deep learning methods and applications. Each model is a collection of explicitly parametrized curves and surfaces, providing ground truth for differential quantities, patch segmentation, geometric feature detection, and shape reconstruction. Sampling the parametric descriptions of surfaces and curves allows generating data in different formats and resolutions, enabling fair comparisons for a wide range of geometric learning algorithms. As a use case for our dataset, we perform a large-scale benchmark for estimation of surface normals, comparing existing data driven methods and evaluating their performance against both the ground truth and traditional normal estimation methods. |
Tasks | |
Published | 2018-12-15 |
URL | http://arxiv.org/abs/1812.06216v2 |
http://arxiv.org/pdf/1812.06216v2.pdf | |
PWC | https://paperswithcode.com/paper/abc-a-big-cad-model-dataset-for-geometric |
Repo | https://github.com/yangyanli/PointCNN |
Framework | tf |