Paper Group ANR 153
Revealing the Hidden Patterns of News Photos: Analysis of Millions of News Photos Using GDELT and Deep Learning-based Vision APIs. Sample Complexity of Automated Mechanism Design. On the performance of different mutation operators of a subpopulation-based genetic algorithm for multi-robot task allocation problems. Polynomial-time Tensor Decompositi …
Revealing the Hidden Patterns of News Photos: Analysis of Millions of News Photos Using GDELT and Deep Learning-based Vision APIs
Title | Revealing the Hidden Patterns of News Photos: Analysis of Millions of News Photos Using GDELT and Deep Learning-based Vision APIs |
Authors | Haewoon Kwak, Jisun An |
Abstract | In this work, we analyze more than two million news photos published in January 2016. We demonstrate i) which objects appear the most in news photos; ii) what the sentiments of news photos are; iii) whether the sentiment of news photos is aligned with the tone of the text; iv) how gender is treated; and v) how differently political candidates are portrayed. To our best knowledge, this is the first large-scale study of news photo contents using deep learning-based vision APIs. |
Tasks | |
Published | 2016-03-15 |
URL | http://arxiv.org/abs/1603.04531v2 |
http://arxiv.org/pdf/1603.04531v2.pdf | |
PWC | https://paperswithcode.com/paper/revealing-the-hidden-patterns-of-news-photos |
Repo | |
Framework | |
Sample Complexity of Automated Mechanism Design
Title | Sample Complexity of Automated Mechanism Design |
Authors | Maria-Florina Balcan, Tuomas Sandholm, Ellen Vitercik |
Abstract | The design of revenue-maximizing combinatorial auctions, i.e. multi-item auctions over bundles of goods, is one of the most fundamental problems in computational economics, unsolved even for two bidders and two items for sale. In the traditional economic models, it is assumed that the bidders’ valuations are drawn from an underlying distribution and that the auction designer has perfect knowledge of this distribution. Despite this strong and oftentimes unrealistic assumption, it is remarkable that the revenue-maximizing combinatorial auction remains unknown. In recent years, automated mechanism design has emerged as one of the most practical and promising approaches to designing high-revenue combinatorial auctions. The most scalable automated mechanism design algorithms take as input samples from the bidders’ valuation distribution and then search for a high-revenue auction in a rich auction class. In this work, we provide the first sample complexity analysis for the standard hierarchy of deterministic combinatorial auction classes used in automated mechanism design. In particular, we provide tight sample complexity bounds on the number of samples needed to guarantee that the empirical revenue of the designed mechanism on the samples is close to its expected revenue on the underlying, unknown distribution over bidder valuations, for each of the auction classes in the hierarchy. In addition to helping set automated mechanism design on firm foundations, our results also push the boundaries of learning theory. In particular, the hypothesis functions used in our contexts are defined through multi-stage combinatorial optimization procedures, rather than simple decision boundaries, as are common in machine learning. |
Tasks | Combinatorial Optimization |
Published | 2016-06-13 |
URL | http://arxiv.org/abs/1606.04145v1 |
http://arxiv.org/pdf/1606.04145v1.pdf | |
PWC | https://paperswithcode.com/paper/sample-complexity-of-automated-mechanism |
Repo | |
Framework | |
On the performance of different mutation operators of a subpopulation-based genetic algorithm for multi-robot task allocation problems
Title | On the performance of different mutation operators of a subpopulation-based genetic algorithm for multi-robot task allocation problems |
Authors | Chun Liu, Andreas Kroll |
Abstract | The performance of different mutation operators is usually evaluated in conjunc-tion with specific parameter settings of genetic algorithms and target problems. Most studies focus on the classical genetic algorithm with different parameters or on solving unconstrained combinatorial optimization problems such as the traveling salesman problems. In this paper, a subpopulation-based genetic al-gorithm that uses only mutation and selection is developed to solve multi-robot task allocation problems. The target problems are constrained combinatorial optimization problems, and are more complex if cooperative tasks are involved as these introduce additional spatial and temporal constraints. The proposed genetic algorithm can obtain better solutions than classical genetic algorithms with tournament selection and partially mapped crossover. The performance of different mutation operators in solving problems without/with cooperative tasks is evaluated. The results imply that inversion mutation performs better than others when solving problems without cooperative tasks, and the swap-inversion combination performs better than others when solving problems with cooperative tasks. |
Tasks | Combinatorial Optimization |
Published | 2016-06-02 |
URL | http://arxiv.org/abs/1606.00601v1 |
http://arxiv.org/pdf/1606.00601v1.pdf | |
PWC | https://paperswithcode.com/paper/on-the-performance-of-different-mutation |
Repo | |
Framework | |
Polynomial-time Tensor Decompositions with Sum-of-Squares
Title | Polynomial-time Tensor Decompositions with Sum-of-Squares |
Authors | Tengyu Ma, Jonathan Shi, David Steurer |
Abstract | We give new algorithms based on the sum-of-squares method for tensor decomposition. Our results improve the best known running times from quasi-polynomial to polynomial for several problems, including decomposing random overcomplete 3-tensors and learning overcomplete dictionaries with constant relative sparsity. We also give the first robust analysis for decomposing overcomplete 4-tensors in the smoothed analysis model. A key ingredient of our analysis is to establish small spectral gaps in moment matrices derived from solutions to sum-of-squares relaxations. To enable this analysis we augment sum-of-squares relaxations with spectral analogs of maximum entropy constraints. |
Tasks | |
Published | 2016-10-06 |
URL | http://arxiv.org/abs/1610.01980v1 |
http://arxiv.org/pdf/1610.01980v1.pdf | |
PWC | https://paperswithcode.com/paper/polynomial-time-tensor-decompositions-with |
Repo | |
Framework | |
Local Perturb-and-MAP for Structured Prediction
Title | Local Perturb-and-MAP for Structured Prediction |
Authors | Gedas Bertasius, Qiang Liu, Lorenzo Torresani, Jianbo Shi |
Abstract | Conditional random fields (CRFs) provide a powerful tool for structured prediction, but cast significant challenges in both the learning and inference steps. Approximation techniques are widely used in both steps, which should be considered jointly to guarantee good performance (a.k.a. “inferning”). Perturb-and-MAP models provide a promising alternative to CRFs, but require global combinatorial optimization and hence they are usable only on specific models. In this work, we present a new Local Perturb-and-MAP (locPMAP) framework that replaces the global optimization with a local optimization by exploiting our observed connection between locPMAP and the pseudolikelihood of the original CRF model. We test our approach on three different vision tasks and show that our method achieves consistently improved performance over other approximate inference techniques optimized to a pseudolikelihood objective. Additionally, we demonstrate that we can integrate our method in the fully convolutional network framework to increase our model’s complexity. Finally, our observed connection between locPMAP and the pseudolikelihood leads to a novel perspective for understanding and using pseudolikelihood. |
Tasks | Combinatorial Optimization, Structured Prediction |
Published | 2016-05-24 |
URL | http://arxiv.org/abs/1605.07686v2 |
http://arxiv.org/pdf/1605.07686v2.pdf | |
PWC | https://paperswithcode.com/paper/local-perturb-and-map-for-structured |
Repo | |
Framework | |
Spectral decomposition method of dialog state tracking via collective matrix factorization
Title | Spectral decomposition method of dialog state tracking via collective matrix factorization |
Authors | Julien Perez |
Abstract | The task of dialog management is commonly decomposed into two sequential subtasks: dialog state tracking and dialog policy learning. In an end-to-end dialog system, the aim of dialog state tracking is to accurately estimate the true dialog state from noisy observations produced by the speech recognition and the natural language understanding modules. The state tracking task is primarily meant to support a dialog policy. From a probabilistic perspective, this is achieved by maintaining a posterior distribution over hidden dialog states composed of a set of context dependent variables. Once a dialog policy is learned, it strives to select an optimal dialog act given the estimated dialog state and a defined reward function. This paper introduces a novel method of dialog state tracking based on a bilinear algebric decomposition model that provides an efficient inference schema through collective matrix factorization. We evaluate the proposed approach on the second Dialog State Tracking Challenge (DSTC-2) dataset and we show that the proposed tracker gives encouraging results compared to the state-of-the-art trackers that participated in this standard benchmark. Finally, we show that the prediction schema is computationally efficient in comparison to the previous approaches. |
Tasks | Speech Recognition |
Published | 2016-06-16 |
URL | http://arxiv.org/abs/1606.05286v1 |
http://arxiv.org/pdf/1606.05286v1.pdf | |
PWC | https://paperswithcode.com/paper/spectral-decomposition-method-of-dialog-state |
Repo | |
Framework | |
Video Processing from Electro-optical Sensors for Object Detection and Tracking in Maritime Environment: A Survey
Title | Video Processing from Electro-optical Sensors for Object Detection and Tracking in Maritime Environment: A Survey |
Authors | D. K. Prasad, D. Rajan, L. Rachmawati, E. Rajabaly, C. Quek |
Abstract | We present a survey on maritime object detection and tracking approaches, which are essential for the development of a navigational system for autonomous ships. The electro-optical (EO) sensor considered here is a video camera that operates in the visible or the infrared spectra, which conventionally complement radar and sonar and have demonstrated effectiveness for situational awareness at sea has demonstrated its effectiveness over the last few years. This paper provides a comprehensive overview of various approaches of video processing for object detection and tracking in the maritime environment. We follow an approach-based taxonomy wherein the advantages and limitations of each approach are compared. The object detection system consists of the following modules: horizon detection, static background subtraction and foreground segmentation. Each of these has been studied extensively in maritime situations and has been shown to be challenging due to the presence of background motion especially due to waves and wakes. The main processes involved in object tracking include video frame registration, dynamic background subtraction, and the object tracking algorithm itself. The challenges for robust tracking arise due to camera motion, dynamic background and low contrast of tracked object, possibly due to environmental degradation. The survey also discusses multisensor approaches and commercial maritime systems that use EO sensors. The survey also highlights methods from computer vision research which hold promise to perform well in maritime EO data processing. Performance of several maritime and computer vision techniques is evaluated on newly proposed Singapore Maritime Dataset. |
Tasks | Object Detection, Object Tracking |
Published | 2016-11-17 |
URL | http://arxiv.org/abs/1611.05842v1 |
http://arxiv.org/pdf/1611.05842v1.pdf | |
PWC | https://paperswithcode.com/paper/video-processing-from-electro-optical-sensors |
Repo | |
Framework | |
Accelerating Deep Convolutional Networks using low-precision and sparsity
Title | Accelerating Deep Convolutional Networks using low-precision and sparsity |
Authors | Ganesh Venkatesh, Eriko Nurvitadhi, Debbie Marr |
Abstract | We explore techniques to significantly improve the compute efficiency and performance of Deep Convolution Networks without impacting their accuracy. To improve the compute efficiency, we focus on achieving high accuracy with extremely low-precision (2-bit) weight networks, and to accelerate the execution time, we aggressively skip operations on zero-values. We achieve the highest reported accuracy of 76.6% Top-1/93% Top-5 on the Imagenet object classification challenge with low-precision network\footnote{github release of the source code coming soon} while reducing the compute requirement by ~3x compared to a full-precision network that achieves similar accuracy. Furthermore, to fully exploit the benefits of our low-precision networks, we build a deep learning accelerator core, dLAC, that can achieve up to 1 TFLOP/mm^2 equivalent for single-precision floating-point operations (~2 TFLOP/mm^2 for half-precision). |
Tasks | Object Classification |
Published | 2016-10-02 |
URL | http://arxiv.org/abs/1610.00324v1 |
http://arxiv.org/pdf/1610.00324v1.pdf | |
PWC | https://paperswithcode.com/paper/accelerating-deep-convolutional-networks |
Repo | |
Framework | |
Geometry in Active Learning for Binary and Multi-class Image Segmentation
Title | Geometry in Active Learning for Binary and Multi-class Image Segmentation |
Authors | Ksenia Konyushkova, Raphael Sznitman, Pascal Fua |
Abstract | We propose an active learning approach to image segmentation that exploits geometric priors to speed up and streamline the annotation process. It can be applied for both background-foreground and multi-class segmentation tasks in 2D images and 3D image volumes. Our approach combines geometric smoothness priors in the image space with more traditional uncertainty measures to estimate which pixels or voxels are the most informative, and thus should to be annotated next. For multi-class settings, we additionally introduce two novel criteria for uncertainty. In the 3D case, we use the resulting uncertainty measure to select voxels lying on a planar patch, which makes batch annotation much more convenient for the end user compared to the setting where voxels are randomly distributed in a volume. The planar patch is found using a branch-and-bound algorithm that looks for a 2D patch in a 3D volume where the most informative instances are located. We evaluate our approach on Electron Microscopy and Magnetic Resonance image volumes, as well as on regular images of horses and faces. We demonstrate a substantial performance increase over other approaches thanks to the use of geometric priors. |
Tasks | Active Learning, Semantic Segmentation |
Published | 2016-06-29 |
URL | http://arxiv.org/abs/1606.09029v4 |
http://arxiv.org/pdf/1606.09029v4.pdf | |
PWC | https://paperswithcode.com/paper/geometry-in-active-learning-for-binary-and |
Repo | |
Framework | |
Automated 5-year Mortality Prediction using Deep Learning and Radiomics Features from Chest Computed Tomography
Title | Automated 5-year Mortality Prediction using Deep Learning and Radiomics Features from Chest Computed Tomography |
Authors | Gustavo Carneiro, Luke Oakden-Rayner, Andrew P. Bradley, Jacinto Nascimento, Lyle Palmer |
Abstract | We propose new methods for the prediction of 5-year mortality in elderly individuals using chest computed tomography (CT). The methods consist of a classifier that performs this prediction using a set of features extracted from the CT image and segmentation maps of multiple anatomic structures. We explore two approaches: 1) a unified framework based on deep learning, where features and classifier are automatically learned in a single optimisation process; and 2) a multi-stage framework based on the design and selection/extraction of hand-crafted radiomics features, followed by the classifier learning process. Experimental results, based on a dataset of 48 annotated chest CTs, show that the deep learning model produces a mean 5-year mortality prediction accuracy of 68.5%, while radiomics produces a mean accuracy that varies between 56% to 66% (depending on the feature selection/extraction method and classifier). The successful development of the proposed models has the potential to make a profound impact in preventive and personalised healthcare. |
Tasks | Computed Tomography (CT), Feature Selection, Mortality Prediction |
Published | 2016-07-01 |
URL | http://arxiv.org/abs/1607.00267v1 |
http://arxiv.org/pdf/1607.00267v1.pdf | |
PWC | https://paperswithcode.com/paper/automated-5-year-mortality-prediction-using |
Repo | |
Framework | |
Face Recognition: Perspectives from the Real-World
Title | Face Recognition: Perspectives from the Real-World |
Authors | Bappaditya Mandal |
Abstract | In this paper, we analyze some of our real-world deployment of face recognition (FR) systems for various applications and discuss the gaps between expectations of the user and what the system can deliver. We evaluate some of our proposed algorithms with ad-hoc modifications for applications such as FR on wearable devices (like Google Glass), monitoring of elderly people in senior citizens centers, FR of children in child care centers and face matching between a scanned IC/passport face image and a few live webcam images for automatic hotel/resort checkouts. We describe each of these applications, the challenges involved and proposed solutions. Since FR is intuitive in nature and we human beings use it for interactions with the outside world, people have high expectations of its performance in real-world scenarios. However, we analyze and discuss here that it is not the case, machine recognition of faces for each of these applications poses unique challenges and demands specific research components so as to adapt in the actual sites. |
Tasks | Face Recognition |
Published | 2016-02-09 |
URL | http://arxiv.org/abs/1602.02999v1 |
http://arxiv.org/pdf/1602.02999v1.pdf | |
PWC | https://paperswithcode.com/paper/face-recognition-perspectives-from-the-real |
Repo | |
Framework | |
A New Training Method for Feedforward Neural Networks Based on Geometric Contraction Property of Activation Functions
Title | A New Training Method for Feedforward Neural Networks Based on Geometric Contraction Property of Activation Functions |
Authors | Petre Birtea, Cosmin Cernazanu-Glavan, Alexandru Sisu |
Abstract | We propose a new training method for a feedforward neural network having the activation functions with the geometric contraction property. The method consists of constructing a new functional that is less nonlinear in comparison with the classical functional by removing the nonlinearity of the activation function from the output layer. We validate this new method by a series of experiments that show an improved learning speed and better classification error. |
Tasks | |
Published | 2016-06-20 |
URL | http://arxiv.org/abs/1606.05990v2 |
http://arxiv.org/pdf/1606.05990v2.pdf | |
PWC | https://paperswithcode.com/paper/a-new-training-method-for-feedforward-neural |
Repo | |
Framework | |
Bayesian Neighbourhood Component Analysis
Title | Bayesian Neighbourhood Component Analysis |
Authors | Dong Wang, Xiaoyang Tan |
Abstract | Learning a good distance metric in feature space potentially improves the performance of the KNN classifier and is useful in many real-world applications. Many metric learning algorithms are however based on the point estimation of a quadratic optimization problem, which is time-consuming, susceptible to overfitting, and lack a natural mechanism to reason with parameter uncertainty, an important property useful especially when the training set is small and/or noisy. To deal with these issues, we present a novel Bayesian metric learning method, called Bayesian NCA, based on the well-known Neighbourhood Component Analysis method, in which the metric posterior is characterized by the local label consistency constraints of observations, encoded with a similarity graph instead of independent pairwise constraints. For efficient Bayesian optimization, we explore the variational lower bound over the log-likelihood of the original NCA objective. Experiments on several publicly available datasets demonstrate that the proposed method is able to learn robust metric measures from small size dataset and/or from challenging training set with labels contaminated by errors. The proposed method is also shown to outperform a previous pairwise constrained Bayesian metric learning method. |
Tasks | Metric Learning |
Published | 2016-04-08 |
URL | http://arxiv.org/abs/1604.02354v1 |
http://arxiv.org/pdf/1604.02354v1.pdf | |
PWC | https://paperswithcode.com/paper/bayesian-neighbourhood-component-analysis |
Repo | |
Framework | |
QPass: a Merit-based Evaluation of Soccer Passes
Title | QPass: a Merit-based Evaluation of Soccer Passes |
Authors | Laszlo Gyarmati, Rade Stanojevic |
Abstract | Quantitative analysis of soccer players’ passing ability focuses on descriptive statistics without considering the players’ real contribution to the passing and ball possession strategy of their team. Which player is able to help the build-up of an attack, or to maintain the possession of the ball? We introduce a novel methodology called QPass to answer questions like these quantitatively. Based on the analysis of an entire season, we rank the players based on the intrinsic value of their passes using QPass. We derive an album of pass trajectories for different gaming styles. Our methodology reveals a quite counterintuitive paradigm: losing the ball possession could lead to better chances to win a game. |
Tasks | |
Published | 2016-08-08 |
URL | http://arxiv.org/abs/1608.03532v1 |
http://arxiv.org/pdf/1608.03532v1.pdf | |
PWC | https://paperswithcode.com/paper/qpass-a-merit-based-evaluation-of-soccer |
Repo | |
Framework | |
Ashwin: Plug-and-Play System for Machine-Human Image Annotation
Title | Ashwin: Plug-and-Play System for Machine-Human Image Annotation |
Authors | Anand Sriraman, Mandar Kulkarni, Rahul Kumar, Kanika Kalra, Purushotam Radadia, Shirish Karande |
Abstract | We present an end-to-end machine-human image annotation system where each component can be attached in a plug-and-play fashion. These components include Feature Extraction, Machine Classifier, Task Sampling and Crowd Consensus. |
Tasks | |
Published | 2016-09-08 |
URL | http://arxiv.org/abs/1609.02271v2 |
http://arxiv.org/pdf/1609.02271v2.pdf | |
PWC | https://paperswithcode.com/paper/ashwin-plug-and-play-system-for-machine-human |
Repo | |
Framework | |