Paper Group ANR 482
Decoding visemes: improving machine lipreading. Accelerating recurrent neural network training using sequence bucketing and multi-GPU data parallelization. A Convolution Tree with Deconvolution Branches: Exploiting Geometric Relationships for Single Shot Keypoint Detection. End-To-End Face Detection and Recognition. Make Your Bone Great Again : A s …
Decoding visemes: improving machine lipreading
Title | Decoding visemes: improving machine lipreading |
Authors | Helen L Bear |
Abstract | Machine lipreading (MLR) is speech recognition from visual cues and a niche research problem in speech processing & computer vision. Current challenges fall into two groups: the content of the video, such as rate of speech or; the parameters of the video recording e.g, video resolution. We show that HD video is not needed to successfully lipread with a computer. The term “viseme” is used in machine lipreading to represent a visual cue or gesture which corresponds to a subgroup of phonemes where the phonemes are visually indistinguishable. A phoneme is the smallest sound one can utter, because there are more phonemes per viseme, maps between units show a many-to-one relationship. Many maps have been presented, we compare these and our results show Lee’s is best. We propose a new method of speaker-dependent phoneme-to-viseme maps and compare these to Lee’s. Our results show the sensitivity of phoneme clustering and we use our new knowledge to augment a conventional MLR system. It has been observed in MLR, that classifiers need training on test subjects to achieve accuracy. Thus machine lipreading is highly speaker-dependent. Conversely speaker independence is robust classification of non-training speakers. We investigate the dependence of phoneme-to-viseme maps between speakers and show there is not a high variability of visemes, but there is high variability in trajectory between visemes of individual speakers with the same ground truth. This implies a dependency upon the number of visemes within each set for each individual. We show that prior phoneme-to-viseme maps rarely have enough visemes and the optimal size, which varies by speaker, ranges from 11-35. Finally we decode from visemes back to phonemes and into words. Our novel approach uses the optimum range visemes within hierarchical training of phoneme classifiers and demonstrates a significant increase in classification accuracy. |
Tasks | Lipreading, Speech Recognition |
Published | 2017-10-03 |
URL | http://arxiv.org/abs/1710.01288v1 |
http://arxiv.org/pdf/1710.01288v1.pdf | |
PWC | https://paperswithcode.com/paper/decoding-visemes-improving-machine-lipreading |
Repo | |
Framework | |
Accelerating recurrent neural network training using sequence bucketing and multi-GPU data parallelization
Title | Accelerating recurrent neural network training using sequence bucketing and multi-GPU data parallelization |
Authors | Viacheslav Khomenko, Oleg Shyshkov, Olga Radyvonenko, Kostiantyn Bokhan |
Abstract | An efficient algorithm for recurrent neural network training is presented. The approach increases the training speed for tasks where a length of the input sequence may vary significantly. The proposed approach is based on the optimal batch bucketing by input sequence length and data parallelization on multiple graphical processing units. The baseline training performance without sequence bucketing is compared with the proposed solution for a different number of buckets. An example is given for the online handwriting recognition task using an LSTM recurrent neural network. The evaluation is performed in terms of the wall clock time, number of epochs, and validation loss value. |
Tasks | |
Published | 2017-08-18 |
URL | http://arxiv.org/abs/1708.05604v1 |
http://arxiv.org/pdf/1708.05604v1.pdf | |
PWC | https://paperswithcode.com/paper/accelerating-recurrent-neural-network |
Repo | |
Framework | |
A Convolution Tree with Deconvolution Branches: Exploiting Geometric Relationships for Single Shot Keypoint Detection
Title | A Convolution Tree with Deconvolution Branches: Exploiting Geometric Relationships for Single Shot Keypoint Detection |
Authors | Amit Kumar, Rama Chellappa |
Abstract | Recently, Deep Convolution Networks (DCNNs) have been applied to the task of face alignment and have shown potential for learning improved feature representations. Although deeper layers can capture abstract concepts like pose, it is difficult to capture the geometric relationships among the keypoints in DCNNs. In this paper, we propose a novel convolution-deconvolution network for facial keypoint detection. Our model predicts the 2D locations of the keypoints and their individual visibility along with 3D head pose, while exploiting the spatial relationships among different keypoints. Different from existing approaches of modeling these relationships, we propose learnable transform functions which captures the relationships between keypoints at feature level. However, due to extensive variations in pose, not all of these relationships act at once, and hence we propose, a pose-based routing function which implicitly models the active relationships. Both transform functions and the routing function are implemented through convolutions in a multi-task framework. Our approach presents a single-shot keypoint detection method, making it different from many existing cascade regression-based methods. We also show that learning these relationships significantly improve the accuracy of keypoint detections for in-the-wild face images from challenging datasets such as AFW and AFLW. |
Tasks | Face Alignment, Keypoint Detection |
Published | 2017-04-06 |
URL | http://arxiv.org/abs/1704.01880v1 |
http://arxiv.org/pdf/1704.01880v1.pdf | |
PWC | https://paperswithcode.com/paper/a-convolution-tree-with-deconvolution |
Repo | |
Framework | |
End-To-End Face Detection and Recognition
Title | End-To-End Face Detection and Recognition |
Authors | Liying Chi, Hongxin Zhang, Mingxiu Chen |
Abstract | Plenty of face detection and recognition methods have been proposed and got delightful results in decades. Common face recognition pipeline consists of: 1) face detection, 2) face alignment, 3) feature extraction, 4) similarity calculation, which are separated and independent from each other. The separated face analyzing stages lead the model redundant calculation and are hard for end-to-end training. In this paper, we proposed a novel end-to-end trainable convolutional network framework for face detection and recognition, in which a geometric transformation matrix was directly learned to align the faces, instead of predicting the facial landmarks. In training stage, our single CNN model is supervised only by face bounding boxes and personal identities, which are publicly available from WIDER FACE \cite{Yang2016} dataset and CASIA-WebFace \cite{Yi2014} dataset. Tested on Face Detection Dataset and Benchmark (FDDB) \cite{Jain2010} dataset and Labeled Face in the Wild (LFW) \cite{Huang2007} dataset, we have achieved 89.24% recall for face detection task and 98.63% verification accuracy for face recognition task simultaneously, which are comparable to state-of-the-art results. |
Tasks | Face Alignment, Face Detection, Face Recognition |
Published | 2017-03-31 |
URL | http://arxiv.org/abs/1703.10818v1 |
http://arxiv.org/pdf/1703.10818v1.pdf | |
PWC | https://paperswithcode.com/paper/end-to-end-face-detection-and-recognition |
Repo | |
Framework | |
Make Your Bone Great Again : A study on Osteoporosis Classification
Title | Make Your Bone Great Again : A study on Osteoporosis Classification |
Authors | Rahul Paul, Saeed Alahamri, Sulav Malla, Ghulam Jilani Quadri |
Abstract | Osteoporosis can be identified by looking at 2D x-ray images of the bone. The high degree of similarity between images of a healthy bone and a diseased one makes classification a challenge. A good bone texture characterization technique is essential for identifying osteoporosis cases. Standard texture feature extraction techniques like Local Binary Pattern (LBP), Gray Level Co-occurrence Matrix (GLCM) have been used for this purpose. In this paper, we draw a comparison between deep features extracted from convolution neural network against these traditional features. Our results show that deep features have more discriminative power as classifiers trained on them always outperform the ones trained on traditional features. |
Tasks | |
Published | 2017-07-17 |
URL | http://arxiv.org/abs/1707.05385v1 |
http://arxiv.org/pdf/1707.05385v1.pdf | |
PWC | https://paperswithcode.com/paper/make-your-bone-great-again-a-study-on |
Repo | |
Framework | |
Gromov-Hausdorff limit of Wasserstein spaces on point clouds
Title | Gromov-Hausdorff limit of Wasserstein spaces on point clouds |
Authors | Nicolas Garcia Trillos |
Abstract | We consider a point cloud $X_n := { x_1, \dots, x_n }$ uniformly distributed on the flat torus $\mathbb{T}^d : = \mathbb{R}^d / \mathbb{Z}^d $, and construct a geometric graph on the cloud by connecting points that are within distance $\varepsilon$ of each other. We let $\mathcal{P}(X_n)$ be the space of probability measures on $X_n$ and endow it with a discrete Wasserstein distance $W_n$ as introduced independently by Chow et al, Maas, and Mielke for general finite Markov chains. We show that as long as $\varepsilon= \varepsilon_n$ decays towards zero slower than an explicit rate depending on the level of uniformity of $X_n$, then the space $(\mathcal{P}(X_n), W_n)$ converges in the Gromov-Hausdorff sense towards the space of probability measures on $\mathbb{T}^d$ endowed with the Wasserstein distance. The analysis presented in this paper is a first step in the study of stability of evolution equations defined over random point clouds as the number of points grows to infinity. |
Tasks | |
Published | 2017-02-11 |
URL | https://arxiv.org/abs/1702.03464v3 |
https://arxiv.org/pdf/1702.03464v3.pdf | |
PWC | https://paperswithcode.com/paper/gromov-hausdorff-limit-of-wasserstein-spaces |
Repo | |
Framework | |
Travel time tomography with adaptive dictionaries
Title | Travel time tomography with adaptive dictionaries |
Authors | Michael Bianco, Peter Gerstoft |
Abstract | We develop a 2D travel time tomography method which regularizes the inversion by modeling groups of slowness pixels from discrete slowness maps, called patches, as sparse linear combinations of atoms from a dictionary. We propose to use dictionary learning during the inversion to adapt dictionaries to specific slowness maps. This patch regularization, called the local model, is integrated into the overall slowness map, called the global model. The local model considers small-scale variations using a sparsity constraint and the global model considers larger-scale features constrained using $\ell_2$ regularization. This strategy in a locally-sparse travel time tomography (LST) approach enables simultaneous modeling of smooth and discontinuous slowness features. This is in contrast to conventional tomography methods, which constrain models to be exclusively smooth or discontinuous. We develop a $\textit{maximum a posteriori}$ formulation for LST and exploit the sparsity of slowness patches using dictionary learning. The LST approach compares favorably with smoothness and total variation regularization methods on densely, but irregularly sampled synthetic slowness maps. |
Tasks | Dictionary Learning |
Published | 2017-12-16 |
URL | http://arxiv.org/abs/1712.08655v3 |
http://arxiv.org/pdf/1712.08655v3.pdf | |
PWC | https://paperswithcode.com/paper/travel-time-tomography-with-adaptive |
Repo | |
Framework | |
GANerated Hands for Real-time 3D Hand Tracking from Monocular RGB
Title | GANerated Hands for Real-time 3D Hand Tracking from Monocular RGB |
Authors | Franziska Mueller, Florian Bernard, Oleksandr Sotnychenko, Dushyant Mehta, Srinath Sridhar, Dan Casas, Christian Theobalt |
Abstract | We address the highly challenging problem of real-time 3D hand tracking based on a monocular RGB-only sequence. Our tracking method combines a convolutional neural network with a kinematic 3D hand model, such that it generalizes well to unseen data, is robust to occlusions and varying camera viewpoints, and leads to anatomically plausible as well as temporally smooth hand motions. For training our CNN we propose a novel approach for the synthetic generation of training data that is based on a geometrically consistent image-to-image translation network. To be more specific, we use a neural network that translates synthetic images to “real” images, such that the so-generated images follow the same statistical distribution as real-world hand images. For training this translation network we combine an adversarial loss and a cycle-consistency loss with a geometric consistency loss in order to preserve geometric properties (such as hand pose) during translation. We demonstrate that our hand tracking system outperforms the current state-of-the-art on challenging RGB-only footage. |
Tasks | Image-to-Image Translation |
Published | 2017-12-04 |
URL | http://arxiv.org/abs/1712.01057v1 |
http://arxiv.org/pdf/1712.01057v1.pdf | |
PWC | https://paperswithcode.com/paper/ganerated-hands-for-real-time-3d-hand |
Repo | |
Framework | |
Comparing Deep Reinforcement Learning and Evolutionary Methods in Continuous Control
Title | Comparing Deep Reinforcement Learning and Evolutionary Methods in Continuous Control |
Authors | Shangtong Zhang, Osmar R. Zaiane |
Abstract | Reinforcement Learning and the Evolutionary Strategy are two major approaches in addressing complicated control problems. Both are strong contenders and have their own devotee communities. Both groups have been very active in developing new advances in their own domain and devising, in recent years, leading-edge techniques to address complex continuous control tasks. Here, in the context of Deep Reinforcement Learning, we formulate a parallelized version of the Proximal Policy Optimization method and a Deep Deterministic Policy Gradient method. Moreover, we conduct a thorough comparison between the state-of-the-art techniques in both camps fro continuous control; evolutionary methods and Deep Reinforcement Learning methods. The results show there is no consistent winner. |
Tasks | Continuous Control |
Published | 2017-11-30 |
URL | http://arxiv.org/abs/1712.00006v2 |
http://arxiv.org/pdf/1712.00006v2.pdf | |
PWC | https://paperswithcode.com/paper/comparing-deep-reinforcement-learning-and |
Repo | |
Framework | |
Learning Pose Grammar to Encode Human Body Configuration for 3D Pose Estimation
Title | Learning Pose Grammar to Encode Human Body Configuration for 3D Pose Estimation |
Authors | Haoshu Fang, Yuanlu Xu, Wenguan Wang, Xiaobai Liu, Song-Chun Zhu |
Abstract | In this paper, we propose a pose grammar to tackle the problem of 3D human pose estimation. Our model directly takes 2D pose as input and learns a generalized 2D-3D mapping function. The proposed model consists of a base network which efficiently captures pose-aligned features and a hierarchy of Bi-directional RNNs (BRNN) on the top to explicitly incorporate a set of knowledge regarding human body configuration (i.e., kinematics, symmetry, motor coordination). The proposed model thus enforces high-level constraints over human poses. In learning, we develop a pose sample simulator to augment training samples in virtual camera views, which further improves our model generalizability. We validate our method on public 3D human pose benchmarks and propose a new evaluation protocol working on cross-view setting to verify the generalization capability of different methods. We empirically observe that most state-of-the-art methods encounter difficulty under such setting while our method can well handle such challenges. |
Tasks | 3D Human Pose Estimation, 3D Pose Estimation, Pose Estimation |
Published | 2017-10-17 |
URL | http://arxiv.org/abs/1710.06513v6 |
http://arxiv.org/pdf/1710.06513v6.pdf | |
PWC | https://paperswithcode.com/paper/learning-pose-grammar-to-encode-human-body |
Repo | |
Framework | |
End-to-end Global to Local CNN Learning for Hand Pose Recovery in Depth Data
Title | End-to-end Global to Local CNN Learning for Hand Pose Recovery in Depth Data |
Authors | Meysam Madadi, Sergio Escalera, Xavier Baro, Jordi Gonzalez |
Abstract | Despite recent advances in 3D pose estimation of human hands, especially thanks to the advent of CNNs and depth cameras, this task is still far from being solved. This is mainly due to the highly non-linear dynamics of fingers, which make hand model training a challenging task. In this paper, we exploit a novel hierarchical tree-like structured CNN, in which branches are trained to become specialized in predefined subsets of hand joints, called local poses. We further fuse local pose features, extracted from hierarchical CNN branches, to learn higher order dependencies among joints in the final pose by end-to-end training. Lastly, the loss function used is also defined to incorporate appearance and physical constraints about doable hand motion and deformation. Finally, we introduce a non-rigid data augmentation approach to increase the amount of training depth data. Experimental results suggest that feeding a tree-shaped CNN, specialized in local poses, into a fusion network for modeling joints correlations and dependencies, helps to increase the precision of final estimations, outperforming state-of-the-art results on NYU and SyntheticHand datasets. |
Tasks | 3D Pose Estimation, Data Augmentation, Pose Estimation |
Published | 2017-05-26 |
URL | http://arxiv.org/abs/1705.09606v2 |
http://arxiv.org/pdf/1705.09606v2.pdf | |
PWC | https://paperswithcode.com/paper/end-to-end-global-to-local-cnn-learning-for |
Repo | |
Framework | |
Beyond opening up the black box: Investigating the role of algorithmic systems in Wikipedian organizational culture
Title | Beyond opening up the black box: Investigating the role of algorithmic systems in Wikipedian organizational culture |
Authors | R. Stuart Geiger |
Abstract | Scholars and practitioners across domains are increasingly concerned with algorithmic transparency and opacity, interrogating the values and assumptions embedded in automated, black-boxed systems, particularly in user-generated content platforms. I report from an ethnography of infrastructure in Wikipedia to discuss an often understudied aspect of this topic: the local, contextual, learned expertise involved in participating in a highly automated social-technical environment. Today, the organizational culture of Wikipedia is deeply intertwined with various data-driven algorithmic systems, which Wikipedians rely on to help manage and govern the “anyone can edit” encyclopedia at a massive scale. These bots, scripts, tools, plugins, and dashboards make Wikipedia more efficient for those who know how to work with them, but like all organizational culture, newcomers must learn them if they want to fully participate. I illustrate how cultural and organizational expertise is enacted around algorithmic agents by discussing two autoethnographic vignettes, which relate my personal experience as a veteran in Wikipedia. I present thick descriptions of how governance and gatekeeping practices are articulated through and in alignment with these automated infrastructures. Over the past 15 years, Wikipedian veterans and administrators have made specific decisions to support administrative and editorial workflows with automation in particular ways and not others. I use these cases of Wikipedia’s bot-supported bureaucracy to discuss several issues in the fields of critical algorithms studies, critical data studies, and fairness, accountability, and transparency in machine learning – most principally arguing that scholarship and practice must go beyond trying to “open up the black box” of such systems and also examine sociocultural processes like newcomer socialization. |
Tasks | |
Published | 2017-09-26 |
URL | http://arxiv.org/abs/1709.09093v2 |
http://arxiv.org/pdf/1709.09093v2.pdf | |
PWC | https://paperswithcode.com/paper/beyond-opening-up-the-black-box-investigating |
Repo | |
Framework | |
Exploring the Combination Rules of D Numbers From a Perspective of Conflict Redistribution
Title | Exploring the Combination Rules of D Numbers From a Perspective of Conflict Redistribution |
Authors | Xinyang Deng, Wen Jiang |
Abstract | Dempster-Shafer theory of evidence is widely applied to uncertainty modelling and knowledge reasoning because of its advantages in dealing with uncertain information. But some conditions or requirements, such as exclusiveness hypothesis and completeness constraint, limit the development and application of that theory to a large extend. To overcome the shortcomings and enhance its capability of representing the uncertainty, a novel model, called D numbers, has been proposed recently. However, many key issues, for example how to implement the combination of D numbers, remain unsolved. In the paper, we have explored the combination of D Numbers from a perspective of conflict redistribution, and proposed two combination rules being suitable for different situations for the fusion of two D numbers. The proposed combination rules can reduce to the classical Dempster’s rule in Dempster-Shafer theory under a certain conditions. Numerical examples and discussion about the proposed rules are also given in the paper. |
Tasks | |
Published | 2017-03-15 |
URL | http://arxiv.org/abs/1703.04862v1 |
http://arxiv.org/pdf/1703.04862v1.pdf | |
PWC | https://paperswithcode.com/paper/exploring-the-combination-rules-of-d-numbers |
Repo | |
Framework | |
Visual gesture variability between talkers in continuous visual speech
Title | Visual gesture variability between talkers in continuous visual speech |
Authors | Helen L Bear |
Abstract | Recent adoption of deep learning methods to the field of machine lipreading research gives us two options to pursue to improve system performance. Either, we develop end-to-end systems holistically or, we experiment to further our understanding of the visual speech signal. The latter option is more difficult but this knowledge would enable researchers to both improve systems and apply the new knowledge to other domains such as speech therapy. One challenge in lipreading systems is the correct labeling of the classifiers. These labels map an estimated function between visemes on the lips and the phonemes uttered. Here we ask if such maps are speaker-dependent? Prior work investigated isolated word recognition from speaker-dependent (SD) visemes, we extend this to continuous speech. Benchmarked against SD results, and the isolated words performance, we test with RMAV dataset speakers and observe that with continuous speech, the trajectory between visemes has a greater negative effect on the speaker differentiation. |
Tasks | Lipreading |
Published | 2017-10-03 |
URL | http://arxiv.org/abs/1710.01297v1 |
http://arxiv.org/pdf/1710.01297v1.pdf | |
PWC | https://paperswithcode.com/paper/visual-gesture-variability-between-talkers-in |
Repo | |
Framework | |
Bootstrapped synthetic likelihood
Title | Bootstrapped synthetic likelihood |
Authors | Richard G. Everitt |
Abstract | Approximate Bayesian computation (ABC) and synthetic likelihood (SL) techniques have enabled the use of Bayesian inference for models that may be simulated, but for which the likelihood cannot be evaluated pointwise at values of an unknown parameter $\theta$. The main idea in ABC and SL is to, for different values of $\theta$ (usually chosen using a Monte Carlo algorithm), build estimates of the likelihood based on simulations from the model conditional on $\theta$. The quality of these estimates determines the efficiency of an ABC/SL algorithm. In standard ABC/SL, the only means to improve an estimated likelihood at $\theta$ is to simulate more times from the model conditional on $\theta$, which is infeasible in cases where the simulator is computationally expensive. In this paper we describe how to use bootstrapping as a means for improving SL estimates whilst using fewer simulations from the model, and also investigate its use in ABC. Further, we investigate the use of the bag of little bootstraps as a means for applying this approach to large datasets, yielding Monte Carlo algorithms that accurately approximate posterior distributions whilst only simulating subsamples of the full data. Examples of the approach applied to i.i.d., temporal and spatial data are given. |
Tasks | Bayesian Inference |
Published | 2017-11-15 |
URL | http://arxiv.org/abs/1711.05825v2 |
http://arxiv.org/pdf/1711.05825v2.pdf | |
PWC | https://paperswithcode.com/paper/bootstrapped-synthetic-likelihood |
Repo | |
Framework | |