Paper Group ANR 227
Robustness Analysis of Pedestrian Detectors for Surveillance. Kid on The Phone! Toward Automatic Detection of Children on Mobile Devices. Synergy: A HW/SW Framework for High Throughput CNNs on Embedded Heterogeneous SoC. Threshold Auto-Tuning Metric Learning. Adversarial Semantic Alignment for Improved Image Captions. Transferable neural networks f …
Robustness Analysis of Pedestrian Detectors for Surveillance
Title | Robustness Analysis of Pedestrian Detectors for Surveillance |
Authors | Yuming Fang, Guanqun Ding, Yuan Yuan, Weisi Lin, Haiwen Liu |
Abstract | To obtain effective pedestrian detection results in surveillance video, there have been many methods proposed to handle the problems from severe occlusion, pose variation, clutter background, \emph{etc}. Besides detection accuracy, a robust surveillance video system should be stable to video quality degradation by network transmission, environment variation, etc. In this study, we conduct the research on the robustness of pedestrian detection algorithms to video quality degradation. The main contribution of this work includes the following three aspects. First, a large-scale Distorted Surveillance Video Data Set (DSurVD) is constructed from high-quality video sequences and their corresponding distorted versions. Second, we design a method to evaluate detection stability and a robustness measure called Robustness Quadrangle, which can be adopted to visualize detection accuracy of pedestrian detection algorithms on high-quality video sequences and stability with video quality degradation. Third, the robustness of seven existing pedestrian detection algorithms is evaluated by the built DSurVD. Experimental results show that the robustness can be further improved for existing pedestrian detection algorithms. Additionally, we provide much in-depth discussion on how different distortion types influence the performance of pedestrian detection algorithms, which is important to design effective pedestrian detection algorithms for surveillance. The DSurVD data set can be download from BaiduYunDisk, https://pan.baidu.com/s/1I9Kqj8rmubOYu7bkBfkUpA, Password: lqmc |
Tasks | Pedestrian Detection |
Published | 2018-07-12 |
URL | http://arxiv.org/abs/1807.04562v2 |
http://arxiv.org/pdf/1807.04562v2.pdf | |
PWC | https://paperswithcode.com/paper/robustness-analysis-of-pedestrian-detectors |
Repo | |
Framework | |
Kid on The Phone! Toward Automatic Detection of Children on Mobile Devices
Title | Kid on The Phone! Toward Automatic Detection of Children on Mobile Devices |
Authors | Toan Nguyen, Aditi Roy, Nasir Memon |
Abstract | Studies have shown that children can be exposed to smart devices at a very early age. This has important implications on research in children-computer interaction, children online safety and early education. Many systems have been built based on such research. In this work, we present multiple techniques to automatically detect the presence of a child on a smart device, which could be used as the first step on such systems. Our methods distinguish children from adults based on behavioral differences while operating a touch-enabled modern computing device. Behavioral differences are extracted from data recorded by the touchscreen and built-in sensors. To evaluate the effectiveness of the proposed methods, a new data set has been created from 50 children and adults who interacted with off-the-shelf applications on smart phones. Results show that it is possible to achieve 99% accuracy and less than 0.5% error rate after 8 consecutive touch gestures using only touch information or 5 seconds of sensor reading. If information is used from multiple sensors, then only after 3 gestures, similar performance could be achieved. |
Tasks | |
Published | 2018-08-05 |
URL | http://arxiv.org/abs/1808.01680v1 |
http://arxiv.org/pdf/1808.01680v1.pdf | |
PWC | https://paperswithcode.com/paper/kid-on-the-phone-toward-automatic-detection |
Repo | |
Framework | |
Synergy: A HW/SW Framework for High Throughput CNNs on Embedded Heterogeneous SoC
Title | Synergy: A HW/SW Framework for High Throughput CNNs on Embedded Heterogeneous SoC |
Authors | Guanwen Zhong, Akshat Dubey, Tan Cheng, Tulika Mitra |
Abstract | Convolutional Neural Networks (CNN) have been widely deployed in diverse application domains. There has been significant progress in accelerating both their training and inference using high-performance GPUs, FPGAs, and custom ASICs for datacenter-scale environments. The recent proliferation of mobile and IoT devices have necessitated real-time, energy-efficient deep neural network inference on embedded-class, resource-constrained platforms. In this context, we present {\em Synergy}, an automated, hardware-software co-designed, pipelined, high-throughput CNN inference framework on embedded heterogeneous system-on-chip (SoC) architectures (Xilinx Zynq). {\em Synergy} leverages, through multi-threading, all the available on-chip resources, which includes the dual-core ARM processor along with the FPGA and the NEON SIMD engines as accelerators. Moreover, {\em Synergy} provides a unified abstraction of the heterogeneous accelerators (FPGA and NEON) and can adapt to different network configurations at runtime without changing the underlying hardware accelerator architecture by balancing workload across accelerators through work-stealing. {\em Synergy} achieves 7.3X speedup, averaged across seven CNN models, over a well-optimized software-only solution. {\em Synergy} demonstrates substantially better throughput and energy-efficiency compared to the contemporary CNN implementations on the same SoC architecture. |
Tasks | |
Published | 2018-03-28 |
URL | http://arxiv.org/abs/1804.00706v1 |
http://arxiv.org/pdf/1804.00706v1.pdf | |
PWC | https://paperswithcode.com/paper/synergy-a-hwsw-framework-for-high-throughput |
Repo | |
Framework | |
Threshold Auto-Tuning Metric Learning
Title | Threshold Auto-Tuning Metric Learning |
Authors | Yuya Onuma, Rachelle Rivero, Tsuyoshi Kato |
Abstract | It has been reported repeatedly that discriminative learning of distance metric boosts the pattern recognition performance. A weak point of ITML-based methods is that the distance threshold for similarity/dissimilarity constraints must be determined manually and it is sensitive to generalization performance, although the ITML-based methods enjoy an advantage that the Bregman projection framework can be applied for optimization of distance metric. In this paper, we present a new formulation of metric learning algorithm in which the distance threshold is optimized together. Since the optimization is still in the Bregman projection framework, the Dykstra algorithm can be applied for optimization. A nonlinear equation has to be solved to project the solution onto a half-space in each iteration. Na"{i}ve method takes $O(LMn^{3})$ computational time to solve the nonlinear equation. In this study, an efficient technique that can solve the nonlinear equation in $O(Mn^{3})$ has been discovered. We have proved that the root exists and is unique. We empirically show that the accuracy of pattern recognition for the proposed metric learning algorithm is comparable to the existing metric learning methods, yet the distance threshold is automatically tuned for the proposed metric learning algorithm. |
Tasks | Metric Learning |
Published | 2018-01-07 |
URL | http://arxiv.org/abs/1801.02125v2 |
http://arxiv.org/pdf/1801.02125v2.pdf | |
PWC | https://paperswithcode.com/paper/threshold-auto-tuning-metric-learning |
Repo | |
Framework | |
Adversarial Semantic Alignment for Improved Image Captions
Title | Adversarial Semantic Alignment for Improved Image Captions |
Authors | Pierre L. Dognin, Igor Melnyk, Youssef Mroueh, Jarret Ross, Tom Sercu |
Abstract | In this paper we study image captioning as a conditional GAN training, proposing both a context-aware LSTM captioner and co-attentive discriminator, which enforces semantic alignment between images and captions. We empirically focus on the viability of two training methods: Self-critical Sequence Training (SCST) and Gumbel Straight-Through (ST) and demonstrate that SCST shows more stable gradient behavior and improved results over Gumbel ST, even without accessing discriminator gradients directly. We also address the problem of automatic evaluation for captioning models and introduce a new semantic score, and show its correlation to human judgement. As an evaluation paradigm, we argue that an important criterion for a captioner is the ability to generalize to compositions of objects that do not usually co-occur together. To this end, we introduce a small captioned Out of Context (OOC) test set. The OOC set, combined with our semantic score, are the proposed new diagnosis tools for the captioning community. When evaluated on OOC and MS-COCO benchmarks, we show that SCST-based training has a strong performance in both semantic score and human evaluation, promising to be a valuable new approach for efficient discrete GAN training. |
Tasks | Image Captioning |
Published | 2018-04-30 |
URL | https://arxiv.org/abs/1805.00063v3 |
https://arxiv.org/pdf/1805.00063v3.pdf | |
PWC | https://paperswithcode.com/paper/improved-image-captioning-with-adversarial |
Repo | |
Framework | |
Transferable neural networks for enhanced sampling of protein dynamics
Title | Transferable neural networks for enhanced sampling of protein dynamics |
Authors | Mohammad M. Sultan, Hannah K. Wayment-Steele, Vijay S. Pande |
Abstract | Variational auto-encoder frameworks have demonstrated success in reducing complex nonlinear dynamics in molecular simulation to a single non-linear embedding. In this work, we illustrate how this non-linear latent embedding can be used as a collective variable for enhanced sampling, and present a simple modification that allows us to rapidly perform sampling in multiple related systems. We first demonstrate our method is able to describe the effects of force field changes in capped alanine dipeptide after learning a model using AMBER99. We further provide a simple extension to variational dynamics encoders that allows the model to be trained in a more efficient manner on larger systems by encoding the outputs of a linear transformation using time-structure based independent component analysis (tICA). Using this technique, we show how such a model trained for one protein, the WW domain, can efficiently be transferred to perform enhanced sampling on a related mutant protein, the GTT mutation. This method shows promise for its ability to rapidly sample related systems using a single transferable collective variable and is generally applicable to sets of related simulations, enabling us to probe the effects of variation in increasingly large systems of biophysical interest. |
Tasks | |
Published | 2018-01-02 |
URL | http://arxiv.org/abs/1801.00636v1 |
http://arxiv.org/pdf/1801.00636v1.pdf | |
PWC | https://paperswithcode.com/paper/transferable-neural-networks-for-enhanced |
Repo | |
Framework | |
Image Reconstruction via Variational Network for Real-Time Hand-Held Sound-Speed Imaging
Title | Image Reconstruction via Variational Network for Real-Time Hand-Held Sound-Speed Imaging |
Authors | Valery Vishnevskiy, Sergio J Sanabria, Orcun Goksel |
Abstract | Speed-of-sound is a biomechanical property for quantitative tissue differentiation, with great potential as a new ultrasound-based image modality. A conventional ultrasound array transducer can be used together with an acoustic mirror, or so-called reflector, to reconstruct sound-speed images from time-of-flight measurements to the reflector collected between transducer element pairs, which constitutes a challenging problem of limited-angle computed tomography. For this problem, we herein present a variational network based image reconstruction architecture that is based on optimization loop unrolling, and provide an efficient training protocol of this network architecture on fully synthetic inclusion data. Our results indicate that the learned model presents good generalization ability, being able to reconstruct images with significantly different statistics compared to the training set. Complex inclusion geometries were shown to be successfully reconstructed, also improving over the prior-art by 23% in reconstruction error and by 10% in contrast on synthetic data. In a phantom study, we demonstrated the detection of multiple inclusions that were not distinguishable by prior-art reconstruction, meanwhile improving the contrast by 27% for a stiff inclusion and by 219% for a soft inclusion. Our reconstruction algorithm takes approximately 10ms, enabling its use as a real-time imaging method on an ultrasound machine, for which we are demonstrating an example preliminary setup herein. |
Tasks | Image Reconstruction |
Published | 2018-07-19 |
URL | http://arxiv.org/abs/1807.07416v1 |
http://arxiv.org/pdf/1807.07416v1.pdf | |
PWC | https://paperswithcode.com/paper/image-reconstruction-via-variational-network |
Repo | |
Framework | |
VirtualIdentity: Privacy-Preserving User Profiling
Title | VirtualIdentity: Privacy-Preserving User Profiling |
Authors | Sisi Wang, Wing-Sea Poon, Golnoosh Farnadi, Caleb Horst, Kebra Thompson, Michael Nickels, Rafael Dowsley, Anderson C. A. Nascimento, Martine De Cock |
Abstract | User profiling from user generated content (UGC) is a common practice that supports the business models of many social media companies. Existing systems require that the UGC is fully exposed to the module that constructs the user profiles. In this paper we show that it is possible to build user profiles without ever accessing the user’s original data, and without exposing the trained machine learning models for user profiling – which are the intellectual property of the company – to the users of the social media site. We present VirtualIdentity, an application that uses secure multi-party cryptographic protocols to detect the age, gender and personality traits of users by classifying their user-generated text and personal pictures with trained support vector machine models in a privacy-preserving manner. |
Tasks | |
Published | 2018-08-30 |
URL | http://arxiv.org/abs/1808.10151v1 |
http://arxiv.org/pdf/1808.10151v1.pdf | |
PWC | https://paperswithcode.com/paper/virtualidentity-privacy-preserving-user |
Repo | |
Framework | |
String Methods for Stochastic Image and Shape Matching
Title | String Methods for Stochastic Image and Shape Matching |
Authors | Alexis Arnaudon, Darryl Holm, Stefan Sommer |
Abstract | Matching of images and analysis of shape differences is traditionally pursued by energy minimization of paths of deformations acting to match the shape objects. In the Large Deformation Diffeomorphic Metric Mapping (LDDMM) framework, iterative gradient descents on the matching functional lead to matching algorithms informally known as Beg algorithms. When stochasticity is introduced to model stochastic variability of shapes and to provide more realistic models of observed shape data, the corresponding matching problem can be solved with a stochastic Beg algorithm, similar to the finite temperature string method used in rare event sampling. In this paper, we apply a stochastic model compatible with the geometry of the LDDMM framework to obtain a stochastic model of images and we derive the stochastic version of the Beg algorithm which we compare with the string method and an expectation-maximization optimization of posterior likelihoods. The algorithm and its use for statistical inference is tested on stochastic LDDMM landmarks and images. |
Tasks | |
Published | 2018-05-15 |
URL | http://arxiv.org/abs/1805.06038v3 |
http://arxiv.org/pdf/1805.06038v3.pdf | |
PWC | https://paperswithcode.com/paper/string-methods-for-stochastic-image-and-shape |
Repo | |
Framework | |
Hardening Deep Neural Networks via Adversarial Model Cascades
Title | Hardening Deep Neural Networks via Adversarial Model Cascades |
Authors | Deepak Vijaykeerthy, Anshuman Suri, Sameep Mehta, Ponnurangam Kumaraguru |
Abstract | Deep neural networks (DNNs) are vulnerable to malicious inputs crafted by an adversary to produce erroneous outputs. Works on securing neural networks against adversarial examples achieve high empirical robustness on simple datasets such as MNIST. However, these techniques are inadequate when empirically tested on complex data sets such as CIFAR-10 and SVHN. Further, existing techniques are designed to target specific attacks and fail to generalize across attacks. We propose the Adversarial Model Cascades (AMC) as a way to tackle the above inadequacies. Our approach trains a cascade of models sequentially where each model is optimized to be robust towards a mixture of multiple attacks. Ultimately, it yields a single model which is secure against a wide range of attacks; namely FGSM, Elastic, Virtual Adversarial Perturbations and Madry. On an average, AMC increases the model’s empirical robustness against various attacks simultaneously, by a significant margin (of 6.225% for MNIST, 5.075% for SVHN and 2.65% for CIFAR10). At the same time, the model’s performance on non-adversarial inputs is comparable to the state-of-the-art models. |
Tasks | |
Published | 2018-02-02 |
URL | http://arxiv.org/abs/1802.01448v4 |
http://arxiv.org/pdf/1802.01448v4.pdf | |
PWC | https://paperswithcode.com/paper/hardening-deep-neural-networks-via |
Repo | |
Framework | |
Acquire, Augment, Segment & Enjoy: Weakly Supervised Instance Segmentation of Supermarket Products
Title | Acquire, Augment, Segment & Enjoy: Weakly Supervised Instance Segmentation of Supermarket Products |
Authors | Patrick Follmann, Bertram Drost, Tobias Böttger |
Abstract | Grocery stores have thousands of products that are usually identified using barcodes with a human in the loop. For automated checkout systems, it is necessary to count and classify the groceries efficiently and robustly. One possibility is to use a deep learning algorithm for instance-aware semantic segmentation. Such methods achieve high accuracies but require a large amount of annotated training data. We propose a system to generate the training annotations in a weakly supervised manner, drastically reducing the labeling effort. We assume that for each training image, only the object class is known. The system automatically segments the corresponding object from the background. The obtained training data is augmented to simulate variations similar to those seen in real-world setups. |
Tasks | Instance Segmentation, Semantic Segmentation, Weakly-supervised instance segmentation |
Published | 2018-07-05 |
URL | http://arxiv.org/abs/1807.02001v2 |
http://arxiv.org/pdf/1807.02001v2.pdf | |
PWC | https://paperswithcode.com/paper/acquire-augment-segment-enjoy-weakly |
Repo | |
Framework | |
Semi-Autoregressive Neural Machine Translation
Title | Semi-Autoregressive Neural Machine Translation |
Authors | Chunqi Wang, Ji Zhang, Haiqing Chen |
Abstract | Existing approaches to neural machine translation are typically autoregressive models. While these models attain state-of-the-art translation quality, they are suffering from low parallelizability and thus slow at decoding long sequences. In this paper, we propose a novel model for fast sequence generation — the semi-autoregressive Transformer (SAT). The SAT keeps the autoregressive property in global but relieves in local and thus is able to produce multiple successive words in parallel at each time step. Experiments conducted on English-German and Chinese-English translation tasks show that the SAT achieves a good balance between translation quality and decoding speed. On WMT’14 English-German translation, the SAT achieves 5.58$\times$ speedup while maintains 88% translation quality, significantly better than the previous non-autoregressive methods. When produces two words at each time step, the SAT is almost lossless (only 1% degeneration in BLEU score). |
Tasks | Machine Translation |
Published | 2018-08-26 |
URL | http://arxiv.org/abs/1808.08583v2 |
http://arxiv.org/pdf/1808.08583v2.pdf | |
PWC | https://paperswithcode.com/paper/semi-autoregressive-neural-machine |
Repo | |
Framework | |
A Multi-Stream Convolutional Neural Network Framework for Group Activity Recognition
Title | A Multi-Stream Convolutional Neural Network Framework for Group Activity Recognition |
Authors | Sina Mokhtarzadeh Azar, Mina Ghadimi Atigh, Ahmad Nickabadi |
Abstract | In this work, we present a framework based on multi-stream convolutional neural networks (CNNs) for group activity recognition. Streams of CNNs are separately trained on different modalities and their predictions are fused at the end. Each stream has two branches to predict the group activity based on person and scene level representations. A new modality based on the human pose estimation is presented to add extra information to the model. We evaluate our method on the Volleyball and Collective Activity datasets. Experimental results show that the proposed framework is able to achieve state-of-the-art results when multiple or single frames are given as input to the model with 90.50% and 86.61% accuracy on Volleyball dataset, respectively, and 87.01% accuracy of multiple frames group activity on Collective Activity dataset. |
Tasks | Activity Recognition, Group Activity Recognition, Pose Estimation |
Published | 2018-12-26 |
URL | http://arxiv.org/abs/1812.10328v1 |
http://arxiv.org/pdf/1812.10328v1.pdf | |
PWC | https://paperswithcode.com/paper/a-multi-stream-convolutional-neural-network |
Repo | |
Framework | |
Fused Deep Neural Networks for Efficient Pedestrian Detection
Title | Fused Deep Neural Networks for Efficient Pedestrian Detection |
Authors | Xianzhi Du, Mostafa El-Khamy, Vlad I. Morariu, Jungwon Lee, Larry Davis |
Abstract | In this paper, we present an efficient pedestrian detection system, designed by fusion of multiple deep neural network (DNN) systems. Pedestrian candidates are first generated by a single shot convolutional multi-box detector at different locations with various scales and aspect ratios. The candidate generator is designed to provide the majority of ground truth pedestrian annotations at the cost of a large number of false positives. Then, a classification system using the idea of ensemble learning is deployed to improve the detection accuracy. The classification system further classifies the generated candidates based on opinions of multiple deep verification networks and a fusion network which utilizes a novel soft-rejection fusion method to adjust the confidence in the detection results. To improve the training of the deep verification networks, a novel soft-label method is devised to assign floating point labels to the generated pedestrian candidates. A deep context aggregation semantic segmentation network also provides pixel-level classification of the scene and its results are softly fused with the detection results by the single shot detector. Our pedestrian detector compared favorably to state-of-art methods on all popular pedestrian detection datasets. For example, our fused DNN has better detection accuracy on the Caltech Pedestrian dataset than all previous state of art methods, while also being the fastest. We significantly improved the log-average miss rate on the Caltech pedestrian dataset to 7.67% and achieved the new state-of-the-art. |
Tasks | Pedestrian Detection, Semantic Segmentation |
Published | 2018-05-02 |
URL | http://arxiv.org/abs/1805.08688v1 |
http://arxiv.org/pdf/1805.08688v1.pdf | |
PWC | https://paperswithcode.com/paper/fused-deep-neural-networks-for-efficient |
Repo | |
Framework | |
Multi-task Learning for Continuous Control
Title | Multi-task Learning for Continuous Control |
Authors | Himani Arora, Rajath Kumar, Jason Krone, Chong Li |
Abstract | Reliable and effective multi-task learning is a prerequisite for the development of robotic agents that can quickly learn to accomplish related, everyday tasks. However, in the reinforcement learning domain, multi-task learning has not exhibited the same level of success as in other domains, such as computer vision. In addition, most reinforcement learning research on multi-task learning has been focused on discrete action spaces, which are not used for robotic control in the real-world. In this work, we apply multi-task learning methods to continuous action spaces and benchmark their performance on a series of simulated continuous control tasks. Most notably, we show that multi-task learning outperforms our baselines and alternative knowledge sharing methods. |
Tasks | Continuous Control, Multi-Task Learning |
Published | 2018-02-03 |
URL | http://arxiv.org/abs/1802.01034v1 |
http://arxiv.org/pdf/1802.01034v1.pdf | |
PWC | https://paperswithcode.com/paper/multi-task-learning-for-continuous-control |
Repo | |
Framework | |