Paper Group AWR 67
Automatic Anomaly Detection in the Cloud Via Statistical Learning. Tick: a Python library for statistical learning, with a particular emphasis on time-dependent modelling. Iterative PET Image Reconstruction Using Convolutional Neural Network Representation. Challenging Neural Dialogue Models with Natural Data: Memory Networks Fail on Incremental Ph …
Automatic Anomaly Detection in the Cloud Via Statistical Learning
Title | Automatic Anomaly Detection in the Cloud Via Statistical Learning |
Authors | Jordan Hochenbaum, Owen S. Vallis, Arun Kejariwal |
Abstract | Performance and high availability have become increasingly important drivers, amongst other drivers, for user retention in the context of web services such as social networks, and web search. Exogenic and/or endogenic factors often give rise to anomalies, making it very challenging to maintain high availability, while also delivering high performance. Given that service-oriented architectures (SOA) typically have a large number of services, with each service having a large set of metrics, automatic detection of anomalies is non-trivial. Although there exists a large body of prior research in anomaly detection, existing techniques are not applicable in the context of social network data, owing to the inherent seasonal and trend components in the time series data. To this end, we developed two novel statistical techniques for automatically detecting anomalies in cloud infrastructure data. Specifically, the techniques employ statistical learning to detect anomalies in both application, and system metrics. Seasonal decomposition is employed to filter the trend and seasonal components of the time series, followed by the use of robust statistical metrics – median and median absolute deviation (MAD) – to accurately detect anomalies, even in the presence of seasonal spikes. We demonstrate the efficacy of the proposed techniques from three different perspectives, viz., capacity planning, user behavior, and supervised learning. In particular, we used production data for evaluation, and we report Precision, Recall, and F-measure in each case. |
Tasks | Anomaly Detection, Time Series |
Published | 2017-04-24 |
URL | http://arxiv.org/abs/1704.07706v1 |
http://arxiv.org/pdf/1704.07706v1.pdf | |
PWC | https://paperswithcode.com/paper/automatic-anomaly-detection-in-the-cloud-via |
Repo | https://github.com/vxld014/S-H-ESD |
Framework | none |
Tick: a Python library for statistical learning, with a particular emphasis on time-dependent modelling
Title | Tick: a Python library for statistical learning, with a particular emphasis on time-dependent modelling |
Authors | Emmanuel Bacry, Martin Bompaire, Stéphane Gaïffas, Soren Poulsen |
Abstract | Tick is a statistical learning library for Python~3, with a particular emphasis on time-dependent models, such as point processes, and tools for generalized linear models and survival analysis. The core of the library is an optimization module providing model computational classes, solvers and proximal operators for regularization. tick relies on a C++ implementation and state-of-the-art optimization algorithms to provide very fast computations in a single node multi-core setting. Source code and documentation can be downloaded from https://github.com/X-DataInitiative/tick |
Tasks | Point Processes, Survival Analysis |
Published | 2017-07-10 |
URL | http://arxiv.org/abs/1707.03003v2 |
http://arxiv.org/pdf/1707.03003v2.pdf | |
PWC | https://paperswithcode.com/paper/tick-a-python-library-for-statistical |
Repo | https://github.com/X-DataInitiative/tick_archive |
Framework | none |
Iterative PET Image Reconstruction Using Convolutional Neural Network Representation
Title | Iterative PET Image Reconstruction Using Convolutional Neural Network Representation |
Authors | Kuang Gong, Jiahui Guan, Kyungsang Kim, Xuezhu Zhang, Georges El Fakhri, Jinyi Qi, Quanzheng Li |
Abstract | PET image reconstruction is challenging due to the ill-poseness of the inverse problem and limited number of detected photons. Recently deep neural networks have been widely and successfully used in computer vision tasks and attracted growing interests in medical imaging. In this work, we trained a deep residual convolutional neural network to improve PET image quality by using the existing inter-patient information. An innovative feature of the proposed method is that we embed the neural network in the iterative reconstruction framework for image representation, rather than using it as a post-processing tool. We formulate the objective function as a constraint optimization problem and solve it using the alternating direction method of multipliers (ADMM) algorithm. Both simulation data and hybrid real data are used to evaluate the proposed method. Quantification results show that our proposed iterative neural network method can outperform the neural network denoising and conventional penalized maximum likelihood methods. |
Tasks | Denoising, Image Reconstruction |
Published | 2017-10-09 |
URL | http://arxiv.org/abs/1710.03344v1 |
http://arxiv.org/pdf/1710.03344v1.pdf | |
PWC | https://paperswithcode.com/paper/iterative-pet-image-reconstruction-using |
Repo | https://github.com/zgongkuang/IterativeCNN |
Framework | tf |
Challenging Neural Dialogue Models with Natural Data: Memory Networks Fail on Incremental Phenomena
Title | Challenging Neural Dialogue Models with Natural Data: Memory Networks Fail on Incremental Phenomena |
Authors | Igor Shalyminov, Arash Eshghi, Oliver Lemon |
Abstract | Natural, spontaneous dialogue proceeds incrementally on a word-by-word basis; and it contains many sorts of disfluency such as mid-utterance/sentence hesitations, interruptions, and self-corrections. But training data for machine learning approaches to dialogue processing is often either cleaned-up or wholly synthetic in order to avoid such phenomena. The question then arises of how well systems trained on such clean data generalise to real spontaneous dialogue, or indeed whether they are trainable at all on naturally occurring dialogue data. To answer this question, we created a new corpus called bAbI+ by systematically adding natural spontaneous incremental dialogue phenomena such as restarts and self-corrections to the Facebook AI Research’s bAbI dialogues dataset. We then explore the performance of a state-of-the-art retrieval model, MemN2N, on this more natural dataset. Results show that the semantic accuracy of the MemN2N model drops drastically; and that although it is in principle able to learn to process the constructions in bAbI+, it needs an impractical amount of training data to do so. Finally, we go on to show that an incremental, semantic parser – DyLan – shows 100% semantic accuracy on both bAbI and bAbI+, highlighting the generalisation properties of linguistically informed dialogue models. |
Tasks | |
Published | 2017-09-22 |
URL | http://arxiv.org/abs/1709.07840v1 |
http://arxiv.org/pdf/1709.07840v1.pdf | |
PWC | https://paperswithcode.com/paper/challenging-neural-dialogue-models-with |
Repo | https://github.com/ishalyminov/babi_tools |
Framework | none |
mixup: Beyond Empirical Risk Minimization
Title | mixup: Beyond Empirical Risk Minimization |
Authors | Hongyi Zhang, Moustapha Cisse, Yann N. Dauphin, David Lopez-Paz |
Abstract | Large deep neural networks are powerful, but exhibit undesirable behaviors such as memorization and sensitivity to adversarial examples. In this work, we propose mixup, a simple learning principle to alleviate these issues. In essence, mixup trains a neural network on convex combinations of pairs of examples and their labels. By doing so, mixup regularizes the neural network to favor simple linear behavior in-between training examples. Our experiments on the ImageNet-2012, CIFAR-10, CIFAR-100, Google commands and UCI datasets show that mixup improves the generalization of state-of-the-art neural network architectures. We also find that mixup reduces the memorization of corrupt labels, increases the robustness to adversarial examples, and stabilizes the training of generative adversarial networks. |
Tasks | Image Classification, Semi-Supervised Image Classification |
Published | 2017-10-25 |
URL | http://arxiv.org/abs/1710.09412v2 |
http://arxiv.org/pdf/1710.09412v2.pdf | |
PWC | https://paperswithcode.com/paper/mixup-beyond-empirical-risk-minimization |
Repo | https://github.com/scut-aitcm/CompetitiveSENet |
Framework | tf |
A Spatiotemporal Oriented Energy Network for Dynamic Texture Recognition
Title | A Spatiotemporal Oriented Energy Network for Dynamic Texture Recognition |
Authors | Isma Hadji, Richard P. Wildes |
Abstract | This paper presents a novel hierarchical spatiotemporal orientation representation for spacetime image analysis. It is designed to combine the benefits of the multilayer architecture of ConvNets and a more controlled approach to spacetime analysis. A distinguishing aspect of the approach is that unlike most contemporary convolutional networks no learning is involved; rather, all design decisions are specified analytically with theoretical motivations. This approach makes it possible to understand what information is being extracted at each stage and layer of processing as well as to minimize heuristic choices in design. Another key aspect of the network is its recurrent nature, whereby the output of each layer of processing feeds back to the input. To keep the network size manageable across layers, a novel cross-channel feature pooling is proposed. The multilayer architecture that results systematically reveals hierarchical image structure in terms of multiscale, multiorientation properties of visual spacetime. To illustrate its utility, the network has been applied to the task of dynamic texture recognition. Empirical evaluation on multiple standard datasets shows that it sets a new state-of-the-art. |
Tasks | Dynamic Texture Recognition |
Published | 2017-08-22 |
URL | http://arxiv.org/abs/1708.06690v1 |
http://arxiv.org/pdf/1708.06690v1.pdf | |
PWC | https://paperswithcode.com/paper/a-spatiotemporal-oriented-energy-network-for |
Repo | https://github.com/hadjisma/soe-net |
Framework | tf |
Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting
Title | Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting |
Authors | Robert Maier, Kihwan Kim, Daniel Cremers, Jan Kautz, Matthias Nießner |
Abstract | We introduce a novel method to obtain high-quality 3D reconstructions from consumer RGB-D sensors. Our core idea is to simultaneously optimize for geometry encoded in a signed distance field (SDF), textures from automatically-selected keyframes, and their camera poses along with material and scene lighting. To this end, we propose a joint surface reconstruction approach that is based on Shape-from-Shading (SfS) techniques and utilizes the estimation of spatially-varying spherical harmonics (SVSH) from subvolumes of the reconstructed scene. Through extensive examples and evaluations, we demonstrate that our method dramatically increases the level of detail in the reconstructed scene geometry and contributes highly to consistent surface texture recovery. |
Tasks | 3D Reconstruction |
Published | 2017-08-04 |
URL | http://arxiv.org/abs/1708.01670v1 |
http://arxiv.org/pdf/1708.01670v1.pdf | |
PWC | https://paperswithcode.com/paper/intrinsic3d-high-quality-3d-reconstruction-by |
Repo | https://github.com/NVlabs/intrinsic3d |
Framework | none |
Learning Intelligent Dialogs for Bounding Box Annotation
Title | Learning Intelligent Dialogs for Bounding Box Annotation |
Authors | Ksenia Konyushkova, Jasper Uijlings, Christoph Lampert, Vittorio Ferrari |
Abstract | We introduce Intelligent Annotation Dialogs for bounding box annotation. We train an agent to automatically choose a sequence of actions for a human annotator to produce a bounding box in a minimal amount of time. Specifically, we consider two actions: box verification, where the annotator verifies a box generated by an object detector, and manual box drawing. We explore two kinds of agents, one based on predicting the probability that a box will be positively verified, and the other based on reinforcement learning. We demonstrate that (1) our agents are able to learn efficient annotation strategies in several scenarios, automatically adapting to the image difficulty, the desired quality of the boxes, and the detector strength; (2) in all scenarios the resulting annotation dialogs speed up annotation compared to manual box drawing alone and box verification alone, while also outperforming any fixed combination of verification and drawing in most scenarios; (3) in a realistic scenario where the detector is iteratively re-trained, our agents evolve a series of strategies that reflect the shifting trade-off between verification and drawing as the detector grows stronger. |
Tasks | |
Published | 2017-12-21 |
URL | http://arxiv.org/abs/1712.08087v3 |
http://arxiv.org/pdf/1712.08087v3.pdf | |
PWC | https://paperswithcode.com/paper/learning-intelligent-dialogs-for-bounding-box |
Repo | https://github.com/google/intelligent_annotation_dialogs |
Framework | tf |
Variational Reasoning for Question Answering with Knowledge Graph
Title | Variational Reasoning for Question Answering with Knowledge Graph |
Authors | Yuyu Zhang, Hanjun Dai, Zornitsa Kozareva, Alexander J. Smola, Le Song |
Abstract | Knowledge graph (KG) is known to be helpful for the task of question answering (QA), since it provides well-structured relational information between entities, and allows one to further infer indirect facts. However, it is challenging to build QA systems which can learn to reason over knowledge graphs based on question-answer pairs alone. First, when people ask questions, their expressions are noisy (for example, typos in texts, or variations in pronunciations), which is non-trivial for the QA system to match those mentioned entities to the knowledge graph. Second, many questions require multi-hop logic reasoning over the knowledge graph to retrieve the answers. To address these challenges, we propose a novel and unified deep learning architecture, and an end-to-end variational learning algorithm which can handle noise in questions, and learn multi-hop reasoning simultaneously. Our method achieves state-of-the-art performance on a recent benchmark dataset in the literature. We also derive a series of new benchmark datasets, including questions for multi-hop reasoning, questions paraphrased by neural translation model, and questions in human voice. Our method yields very promising results on all these challenging datasets. |
Tasks | Knowledge Graphs, Question Answering |
Published | 2017-09-12 |
URL | http://arxiv.org/abs/1709.04071v5 |
http://arxiv.org/pdf/1709.04071v5.pdf | |
PWC | https://paperswithcode.com/paper/variational-reasoning-for-question-answering |
Repo | https://github.com/yuyuz/MetaQA |
Framework | none |
Neural Factorization Machines for Sparse Predictive Analytics
Title | Neural Factorization Machines for Sparse Predictive Analytics |
Authors | Xiangnan He, Tat-Seng Chua |
Abstract | Many predictive tasks of web applications need to model categorical variables, such as user IDs and demographics like genders and occupations. To apply standard machine learning techniques, these categorical predictors are always converted to a set of binary features via one-hot encoding, making the resultant feature vector highly sparse. To learn from such sparse data effectively, it is crucial to account for the interactions between features. Factorization Machines (FMs) are a popular solution for efficiently using the second-order feature interactions. However, FM models feature interactions in a linear way, which can be insufficient for capturing the non-linear and complex inherent structure of real-world data. While deep neural networks have recently been applied to learn non-linear feature interactions in industry, such as the Wide&Deep by Google and DeepCross by Microsoft, the deep structure meanwhile makes them difficult to train. In this paper, we propose a novel model Neural Factorization Machine (NFM) for prediction under sparse settings. NFM seamlessly combines the linearity of FM in modelling second-order feature interactions and the non-linearity of neural network in modelling higher-order feature interactions. Conceptually, NFM is more expressive than FM since FM can be seen as a special case of NFM without hidden layers. Empirical results on two regression tasks show that with one hidden layer only, NFM significantly outperforms FM with a 7.3% relative improvement. Compared to the recent deep learning methods Wide&Deep and DeepCross, our NFM uses a shallower structure but offers better performance, being much easier to train and tune in practice. |
Tasks | |
Published | 2017-08-16 |
URL | http://arxiv.org/abs/1708.05027v1 |
http://arxiv.org/pdf/1708.05027v1.pdf | |
PWC | https://paperswithcode.com/paper/neural-factorization-machines-for-sparse |
Repo | https://github.com/shenweichen/DeepCTR |
Framework | tf |
Tacotron: Towards End-to-End Speech Synthesis
Title | Tacotron: Towards End-to-End Speech Synthesis |
Authors | Yuxuan Wang, RJ Skerry-Ryan, Daisy Stanton, Yonghui Wu, Ron J. Weiss, Navdeep Jaitly, Zongheng Yang, Ying Xiao, Zhifeng Chen, Samy Bengio, Quoc Le, Yannis Agiomyrgiannakis, Rob Clark, Rif A. Saurous |
Abstract | A text-to-speech synthesis system typically consists of multiple stages, such as a text analysis frontend, an acoustic model and an audio synthesis module. Building these components often requires extensive domain expertise and may contain brittle design choices. In this paper, we present Tacotron, an end-to-end generative text-to-speech model that synthesizes speech directly from characters. Given <text, audio> pairs, the model can be trained completely from scratch with random initialization. We present several key techniques to make the sequence-to-sequence framework perform well for this challenging task. Tacotron achieves a 3.82 subjective 5-scale mean opinion score on US English, outperforming a production parametric system in terms of naturalness. In addition, since Tacotron generates speech at the frame level, it’s substantially faster than sample-level autoregressive methods. |
Tasks | Speech Synthesis, Text-To-Speech Synthesis |
Published | 2017-03-29 |
URL | http://arxiv.org/abs/1703.10135v2 |
http://arxiv.org/pdf/1703.10135v2.pdf | |
PWC | https://paperswithcode.com/paper/tacotron-towards-end-to-end-speech-synthesis |
Repo | https://github.com/0fengzi0/tacotron |
Framework | tf |
Fast Exact k-Means, k-Medians and Bregman Divergence Clustering in 1D
Title | Fast Exact k-Means, k-Medians and Bregman Divergence Clustering in 1D |
Authors | Allan Grønlund, Kasper Green Larsen, Alexander Mathiasen, Jesper Sindahl Nielsen, Stefan Schneider, Mingzhou Song |
Abstract | The $k$-Means clustering problem on $n$ points is NP-Hard for any dimension $d\ge 2$, however, for the 1D case there exists exact polynomial time algorithms. Previous literature reported an $O(kn^2)$ time dynamic programming algorithm that uses $O(kn)$ space. It turns out that the problem has been considered under a different name more than twenty years ago. We present all the existing work that had been overlooked and compare the various solutions theoretically. Moreover, we show how to reduce the space usage for some of them, as well as generalize them to data structures that can quickly report an optimal $k$-Means clustering for any $k$. Finally we also generalize all the algorithms to work for the absolute distance and to work for any Bregman Divergence. We complement our theoretical contributions by experiments that compare the practical performance of the various algorithms. |
Tasks | |
Published | 2017-01-25 |
URL | http://arxiv.org/abs/1701.07204v4 |
http://arxiv.org/pdf/1701.07204v4.pdf | |
PWC | https://paperswithcode.com/paper/fast-exact-k-means-k-medians-and-bregman |
Repo | https://github.com/dstein64/kmeans1d |
Framework | none |
Dilated Recurrent Neural Networks
Title | Dilated Recurrent Neural Networks |
Authors | Shiyu Chang, Yang Zhang, Wei Han, Mo Yu, Xiaoxiao Guo, Wei Tan, Xiaodong Cui, Michael Witbrock, Mark Hasegawa-Johnson, Thomas S. Huang |
Abstract | Learning with recurrent neural networks (RNNs) on long sequences is a notoriously difficult task. There are three major challenges: 1) complex dependencies, 2) vanishing and exploding gradients, and 3) efficient parallelization. In this paper, we introduce a simple yet effective RNN connection structure, the DilatedRNN, which simultaneously tackles all of these challenges. The proposed architecture is characterized by multi-resolution dilated recurrent skip connections and can be combined flexibly with diverse RNN cells. Moreover, the DilatedRNN reduces the number of parameters needed and enhances training efficiency significantly, while matching state-of-the-art performance (even with standard RNN cells) in tasks involving very long-term dependencies. To provide a theory-based quantification of the architecture’s advantages, we introduce a memory capacity measure, the mean recurrent length, which is more suitable for RNNs with long skip connections than existing measures. We rigorously prove the advantages of the DilatedRNN over other recurrent neural architectures. The code for our method is publicly available at https://github.com/code-terminator/DilatedRNN |
Tasks | Sequential Image Classification |
Published | 2017-10-05 |
URL | http://arxiv.org/abs/1710.02224v3 |
http://arxiv.org/pdf/1710.02224v3.pdf | |
PWC | https://paperswithcode.com/paper/dilated-recurrent-neural-networks |
Repo | https://github.com/code-terminator/DilatedRNN |
Framework | tf |
The Minor Fall, the Major Lift: Inferring Emotional Valence of Musical Chords through Lyrics
Title | The Minor Fall, the Major Lift: Inferring Emotional Valence of Musical Chords through Lyrics |
Authors | Artemy Kolchinsky, Nakul Dhande, Kengjeun Park, Yong-Yeol Ahn |
Abstract | We investigate the association between musical chords and lyrics by analyzing a large dataset of user-contributed guitar tablatures. Motivated by the idea that the emotional content of chords is reflected in the words used in corresponding lyrics, we analyze associations between lyrics and chord categories. We also examine the usage patterns of chords and lyrics in different musical genres, historical eras, and geographical regions. Our overall results confirms a previously known association between Major chords and positive valence. We also report a wide variation in this association across regions, genres, and eras. Our results suggest possible existence of different emotional associations for other types of chords. |
Tasks | |
Published | 2017-06-26 |
URL | http://arxiv.org/abs/1706.08609v2 |
http://arxiv.org/pdf/1706.08609v2.pdf | |
PWC | https://paperswithcode.com/paper/the-minor-fall-the-major-lift-inferring |
Repo | https://github.com/artemyk/chordsentiment |
Framework | none |
Sparse quadratic classification rules via linear dimension reduction
Title | Sparse quadratic classification rules via linear dimension reduction |
Authors | Irina Gaynanova, Tianying Wang |
Abstract | We consider the problem of high-dimensional classification between the two groups with unequal covariance matrices. Rather than estimating the full quadratic discriminant rule, we propose to perform simultaneous variable selection and linear dimension reduction on original data, with the subsequent application of quadratic discriminant analysis on the reduced space. In contrast to quadratic discriminant analysis, the proposed framework doesn’t require estimation of precision matrices and scales linearly with the number of measurements, making it especially attractive for the use on high-dimensional datasets. We support the methodology with theoretical guarantees on variable selection consistency, and empirical comparison with competing approaches. We apply the method to gene expression data of breast cancer patients, and confirm the crucial importance of ESR1 gene in differentiating estrogen receptor status. |
Tasks | Dimensionality Reduction |
Published | 2017-11-13 |
URL | http://arxiv.org/abs/1711.04817v2 |
http://arxiv.org/pdf/1711.04817v2.pdf | |
PWC | https://paperswithcode.com/paper/sparse-quadratic-classification-rules-via |
Repo | https://github.com/irinagain/DAP |
Framework | none |