Paper Group ANR 1141
A Simple Quantum Neural Net with a Periodic Activation Function. 30m resolution Global Annual Burned Area Mapping based on Landsat images and Google Earth Engine. A Survey on Semantic Parsing. Universal Decision-Based Black-Box Perturbations: Breaking Security-Through-Obscurity Defenses. Cross Script Hindi English NER Corpus from Wikipedia. Augment …
A Simple Quantum Neural Net with a Periodic Activation Function
Title | A Simple Quantum Neural Net with a Periodic Activation Function |
Authors | Ammar Daskin |
Abstract | In this paper, we propose a simple neural net that requires only $O(nlog_2k)$ number of qubits and $O(nk)$ quantum gates: Here, $n$ is the number of input parameters, and $k$ is the number of weights applied to these parameters in the proposed neural net. We describe the network in terms of a quantum circuit, and then draw its equivalent classical neural net which involves $O(k^n)$ nodes in the hidden layer. Then, we show that the network uses a periodic activation function of cosine values of the linear combinations of the inputs and weights. The backpropagation is described through the gradient descent, and then iris and breast cancer datasets are used for the simulations. The numerical results indicate the network can be used in machine learning problems and it may provide exponential speedup over the same structured classical neural net. |
Tasks | |
Published | 2018-04-20 |
URL | http://arxiv.org/abs/1804.07633v3 |
http://arxiv.org/pdf/1804.07633v3.pdf | |
PWC | https://paperswithcode.com/paper/a-simple-quantum-neural-net-with-a-periodic |
Repo | |
Framework | |
30m resolution Global Annual Burned Area Mapping based on Landsat images and Google Earth Engine
Title | 30m resolution Global Annual Burned Area Mapping based on Landsat images and Google Earth Engine |
Authors | Tengfei Long, Zhaoming Zhang, Guojin He, Weili Jiao, Chao Tang, Bingfang Wu, Xiaomei Zhang, Guizhou Wang, Ranyu Yin |
Abstract | Heretofore, global burned area (BA) products are only available at coarse spatial resolution, since most of the current global BA products are produced with the help of active fire detection or dense time-series change analysis, which requires very high temporal resolution. In this study, however, we focus on automated global burned area mapping approach based on Landsat images. By utilizing the huge catalog of satellite imagery as well as the high-performance computing capacity of Google Earth Engine, we proposed an automated pipeline for generating 30-meter resolution global-scale annual burned area map from time-series of Landsat images, and a novel 30-meter resolution global annual burned area map of 2015 (GABAM 2015) is released. GABAM 2015 consists of spatial extent of fires that occurred during 2015 and not of fires that occurred in previous years. Cross-comparison with recent Fire_cci version 5.0 BA product found a similar spatial distribution and a strong correlation ($R^2=0.74$) between the burned areas from the two products, although differences were found in specific land cover categories (particularly in agriculture land). Preliminary global validation showed the commission and omission error of GABAM 2015 are 13.17% and 30.13%, respectively. |
Tasks | Time Series |
Published | 2018-05-07 |
URL | http://arxiv.org/abs/1805.02579v1 |
http://arxiv.org/pdf/1805.02579v1.pdf | |
PWC | https://paperswithcode.com/paper/30m-resolution-global-annual-burned-area |
Repo | |
Framework | |
A Survey on Semantic Parsing
Title | A Survey on Semantic Parsing |
Authors | Aishwarya Kamath, Rajarshi Das |
Abstract | A significant amount of information in today’s world is stored in structured and semi-structured knowledge bases. Efficient and simple methods to query them are essential and must not be restricted to only those who have expertise in formal query languages. The field of semantic parsing deals with converting natural language utterances to logical forms that can be easily executed on a knowledge base. In this survey, we examine the various components of a semantic parsing system and discuss prominent work ranging from the initial rule based methods to the current neural approaches to program synthesis. We also discuss methods that operate using varying levels of supervision and highlight the key challenges involved in the learning of such systems. |
Tasks | Program Synthesis, Semantic Parsing |
Published | 2018-12-03 |
URL | https://arxiv.org/abs/1812.00978v3 |
https://arxiv.org/pdf/1812.00978v3.pdf | |
PWC | https://paperswithcode.com/paper/a-survey-on-semantic-parsing |
Repo | |
Framework | |
Universal Decision-Based Black-Box Perturbations: Breaking Security-Through-Obscurity Defenses
Title | Universal Decision-Based Black-Box Perturbations: Breaking Security-Through-Obscurity Defenses |
Authors | Thomas A. Hogan, Bhavya Kailkhura |
Abstract | We study the problem of finding a universal (image-agnostic) perturbation to fool machine learning (ML) classifiers (e.g., neural nets, decision tress) in the hard-label black-box setting. Recent work in adversarial ML in the white-box setting (model parameters are known) has shown that many state-of-the-art image classifiers are vulnerable to universal adversarial perturbations: a fixed human-imperceptible perturbation that, when added to any image, causes it to be misclassified with high probability Kurakin et al. [2016], Szegedy et al. [2013], Chen et al. [2017a], Carlini and Wagner [2017]. This paper considers a more practical and challenging problem of finding such universal perturbations in an obscure (or black-box) setting. More specifically, we use zeroth order optimization algorithms to find such a universal adversarial perturbation when no model information is revealed-except that the attacker can make queries to probe the classifier. We further relax the assumption that the output of a query is continuous valued confidence scores for all the classes and consider the case where the output is a hard-label decision. Surprisingly, we found that even in these extremely obscure regimes, state-of-the-art ML classifiers can be fooled with a very high probability just by adding a single human-imperceptible image perturbation to any natural image. The surprising existence of universal perturbations in a hard-label black-box setting raises serious security concerns with the existence of a universal noise vector that adversaries can possibly exploit to break a classifier on most natural images. |
Tasks | |
Published | 2018-11-09 |
URL | http://arxiv.org/abs/1811.03733v2 |
http://arxiv.org/pdf/1811.03733v2.pdf | |
PWC | https://paperswithcode.com/paper/universal-decision-based-black-box |
Repo | |
Framework | |
Cross Script Hindi English NER Corpus from Wikipedia
Title | Cross Script Hindi English NER Corpus from Wikipedia |
Authors | Mohd Zeeshan Ansari, Tanvir Ahmad, Md Arshad Ali |
Abstract | The text generated on social media platforms is essentially a mixed lingual text. The mixing of language in any form produces considerable amount of difficulty in language processing systems. Moreover, the advancements in language processing research depends upon the availability of standard corpora. The development of mixed lingual Indian Named Entity Recognition (NER) systems are facing obstacles due to unavailability of the standard evaluation corpora. Such corpora may be of mixed lingual nature in which text is written using multiple languages predominantly using a single script only. The motivation of our work is to emphasize the automatic generation such kind of corpora in order to encourage mixed lingual Indian NER. The paper presents the preparation of a Cross Script Hindi-English Corpora from Wikipedia category pages. The corpora is successfully annotated using standard CoNLL-2003 categories of PER, LOC, ORG, and MISC. Its evaluation is carried out on a variety of machine learning algorithms and favorable results are achieved. |
Tasks | Named Entity Recognition |
Published | 2018-10-08 |
URL | http://arxiv.org/abs/1810.03430v1 |
http://arxiv.org/pdf/1810.03430v1.pdf | |
PWC | https://paperswithcode.com/paper/cross-script-hindi-english-ner-corpus-from |
Repo | |
Framework | |
Augmented Skeleton Space Transfer for Depth-based Hand Pose Estimation
Title | Augmented Skeleton Space Transfer for Depth-based Hand Pose Estimation |
Authors | Seungryul Baek, Kwang In Kim, Tae-Kyun Kim |
Abstract | Crucial to the success of training a depth-based 3D hand pose estimator (HPE) is the availability of comprehensive datasets covering diverse camera perspectives, shapes, and pose variations. However, collecting such annotated datasets is challenging. We propose to complete existing databases by generating new database entries. The key idea is to synthesize data in the skeleton space (instead of doing so in the depth-map space) which enables an easy and intuitive way of manipulating data entries. Since the skeleton entries generated in this way do not have the corresponding depth map entries, we exploit them by training a separate hand pose generator (HPG) which synthesizes the depth map from the skeleton entries. By training the HPG and HPE in a single unified optimization framework enforcing that 1) the HPE agrees with the paired depth and skeleton entries; and 2) the HPG-HPE combination satisfies the cyclic consistency (both the input and the output of HPG-HPE are skeletons) observed via the newly generated unpaired skeletons, our algorithm constructs a HPE which is robust to variations that go beyond the coverage of the existing database. Our training algorithm adopts the generative adversarial networks (GAN) training process. As a by-product, we obtain a hand pose discriminator (HPD) that is capable of picking out realistic hand poses. Our algorithm exploits this capability to refine the initial skeleton estimates in testing, further improving the accuracy. We test our algorithm on four challenging benchmark datasets (ICVL, MSRA, NYU and Big Hand 2.2M datasets) and demonstrate that our approach outperforms or is on par with state-of-the-art methods quantitatively and qualitatively. |
Tasks | Hand Pose Estimation, Pose Estimation |
Published | 2018-05-11 |
URL | http://arxiv.org/abs/1805.04497v1 |
http://arxiv.org/pdf/1805.04497v1.pdf | |
PWC | https://paperswithcode.com/paper/augmented-skeleton-space-transfer-for-depth |
Repo | |
Framework | |
Learning Pixel-level Semantic Affinity with Image-level Supervision for Weakly Supervised Semantic Segmentation
Title | Learning Pixel-level Semantic Affinity with Image-level Supervision for Weakly Supervised Semantic Segmentation |
Authors | Jiwoon Ahn, Suha Kwak |
Abstract | The deficiency of segmentation labels is one of the main obstacles to semantic segmentation in the wild. To alleviate this issue, we present a novel framework that generates segmentation labels of images given their image-level class labels. In this weakly supervised setting, trained models have been known to segment local discriminative parts rather than the entire object area. Our solution is to propagate such local responses to nearby areas which belong to the same semantic entity. To this end, we propose a Deep Neural Network (DNN) called AffinityNet that predicts semantic affinity between a pair of adjacent image coordinates. The semantic propagation is then realized by random walk with the affinities predicted by AffinityNet. More importantly, the supervision employed to train AffinityNet is given by the initial discriminative part segmentation, which is incomplete as a segmentation annotation but sufficient for learning semantic affinities within small image areas. Thus the entire framework relies only on image-level class labels and does not require any extra data or annotations. On the PASCAL VOC 2012 dataset, a DNN learned with segmentation labels generated by our method outperforms previous models trained with the same level of supervision, and is even as competitive as those relying on stronger supervision. |
Tasks | Semantic Segmentation, Weakly-Supervised Semantic Segmentation |
Published | 2018-03-28 |
URL | http://arxiv.org/abs/1803.10464v2 |
http://arxiv.org/pdf/1803.10464v2.pdf | |
PWC | https://paperswithcode.com/paper/learning-pixel-level-semantic-affinity-with |
Repo | |
Framework | |
Articulatory information and Multiview Features for Large Vocabulary Continuous Speech Recognition
Title | Articulatory information and Multiview Features for Large Vocabulary Continuous Speech Recognition |
Authors | Vikramjit Mitra, Wen Wang, Chris Bartels, Horacio Franco, Dimitra Vergyri |
Abstract | This paper explores the use of multi-view features and their discriminative transforms in a convolutional deep neural network (CNN) architecture for a continuous large vocabulary speech recognition task. Mel-filterbank energies and perceptually motivated forced damped oscillator coefficient (DOC) features are used after feature-space maximum-likelihood linear regression (fMLLR) transforms, which are combined and fed as a multi-view feature to a single CNN acoustic model. Use of multi-view feature representation demonstrated significant reduction in word error rates (WERs) compared to the use of individual features by themselves. In addition, when articulatory information was used as an additional input to a fused deep neural network (DNN) and CNN acoustic model, it was found to demonstrate further reduction in WER for the Switchboard subset and the CallHome subset (containing partly non-native accented speech) of the NIST 2000 conversational telephone speech test set, reducing the error rate by 12% relative to the baseline in both cases. This work shows that multi-view features in association with articulatory information can improve speech recognition robustness to spontaneous and non-native speech. |
Tasks | Large Vocabulary Continuous Speech Recognition, Speech Recognition |
Published | 2018-02-16 |
URL | http://arxiv.org/abs/1802.05853v1 |
http://arxiv.org/pdf/1802.05853v1.pdf | |
PWC | https://paperswithcode.com/paper/articulatory-information-and-multiview |
Repo | |
Framework | |
Sparse component separation from Poisson measurements
Title | Sparse component separation from Poisson measurements |
Authors | I. El Hamzaoui, J. Bobin |
Abstract | Blind source separation (BSS) aims at recovering signals from mixtures. This problem has been extensively studied in cases where the mixtures are contaminated with additive Gaussian noise. However, it is not well suited to describe data that are corrupted with Poisson measurements such as in low photon count optics or in high-energy astronomical imaging (e.g. observations from the Chandra or Fermi telescopes). To that purpose, we propose a novel BSS algorithm coined pGMCA that specifically tackles the blind separation of sparse sources from Poisson measurements. |
Tasks | |
Published | 2018-12-11 |
URL | http://arxiv.org/abs/1812.04370v1 |
http://arxiv.org/pdf/1812.04370v1.pdf | |
PWC | https://paperswithcode.com/paper/sparse-component-separation-from-poisson |
Repo | |
Framework | |
An LP-based hyperparameter optimization model for language modeling
Title | An LP-based hyperparameter optimization model for language modeling |
Authors | Amir Hossein Akhavan Rahnama, Mehdi Toloo, Nezer Jacob Zaidenberg |
Abstract | In order to find hyperparameters for a machine learning model, algorithms such as grid search or random search are used over the space of possible values of the models hyperparameters. These search algorithms opt the solution that minimizes a specific cost function. In language models, perplexity is one of the most popular cost functions. In this study, we propose a fractional nonlinear programming model that finds the optimal perplexity value. The special structure of the model allows us to approximate it by a linear programming model that can be solved using the well-known simplex algorithm. To the best of our knowledge, this is the first attempt to use optimization techniques to find perplexity values in the language modeling literature. We apply our model to find hyperparameters of a language model and compare it to the grid search algorithm. Furthermore, we illustrating that it results in lower perplexity values. We perform this experiment on a real-world dataset from SwiftKey to validate our proposed approach. |
Tasks | Hyperparameter Optimization, Language Modelling |
Published | 2018-03-29 |
URL | http://arxiv.org/abs/1803.10927v1 |
http://arxiv.org/pdf/1803.10927v1.pdf | |
PWC | https://paperswithcode.com/paper/an-lp-based-hyperparameter-optimization-model |
Repo | |
Framework | |
Structured low-rank matrix completion for forecasting in time series analysis
Title | Structured low-rank matrix completion for forecasting in time series analysis |
Authors | Jonathan Gillard, Konstantin Usevich |
Abstract | In this paper we consider the low-rank matrix completion problem with specific application to forecasting in time series analysis. Briefly, the low-rank matrix completion problem is the problem of imputing missing values of a matrix under a rank constraint. We consider a matrix completion problem for Hankel matrices and a convex relaxation based on the nuclear norm. Based on new theoretical results and a number of numerical and real examples, we investigate the cases when the proposed approach can work. Our results highlight the importance of choosing a proper weighting scheme for the known observations. |
Tasks | Low-Rank Matrix Completion, Matrix Completion, Time Series, Time Series Analysis |
Published | 2018-02-22 |
URL | http://arxiv.org/abs/1802.08242v1 |
http://arxiv.org/pdf/1802.08242v1.pdf | |
PWC | https://paperswithcode.com/paper/structured-low-rank-matrix-completion-for |
Repo | |
Framework | |
MsCGAN: Multi-scale Conditional Generative Adversarial Networks for Person Image Generation
Title | MsCGAN: Multi-scale Conditional Generative Adversarial Networks for Person Image Generation |
Authors | Wei Tang, Gui Li, Xinyuan Bao, Teng Li |
Abstract | To synthesize high-quality person images with arbitrary poses is challenging. In this paper, we propose a novel Multi-scale Conditional Generative Adversarial Networks (MsCGAN), aiming to convert the input conditional person image to a synthetic image of any given target pose, whose appearance and the texture are consistent with the input image. MsCGAN is a multi-scale adversarial network consisting of two generators and two discriminators. One generator transforms the conditional person image into a coarse image of the target pose globally, and the other is to enhance the detailed quality of the synthetic person image through a local reinforcement network. The outputs of the two generators are then merged into a synthetic, discriminant and high-resolution image. On the other hand, the synthetic image is downsampled to multiple resolutions as the input to multi-scale discriminator networks. The proposed multi-scale generators and discriminators handling different levels of visual features can benefit to synthesizing high-resolution person images with realistic appearance and texture. Experiments are conducted on the Market-1501 and DeepFashion datasets to evaluate the proposed model, and both qualitative and quantitative results demonstrate the superior performance of the proposed MsCGAN. |
Tasks | Image Generation |
Published | 2018-10-19 |
URL | https://arxiv.org/abs/1810.08534v3 |
https://arxiv.org/pdf/1810.08534v3.pdf | |
PWC | https://paperswithcode.com/paper/mscgan-multi-scale-conditional-generative |
Repo | |
Framework | |
A Mathematical Framework for Superintelligent Machines
Title | A Mathematical Framework for Superintelligent Machines |
Authors | Daniel J. Buehrer |
Abstract | We describe a class calculus that is expressive enough to describe and improve its own learning process. It can design and debug programs that satisfy given input/output constraints, based on its ontology of previously learned programs. It can improve its own model of the world by checking the actual results of the actions of its robotic activators. For instance, it could check the black box of a car crash to determine if it was probably caused by electric failure, a stuck electronic gate, dark ice, or some other condition that it must add to its ontology in order to meet its sub-goal of preventing such crashes in the future. Class algebra basically defines the eval/eval-1 Galois connection between the residuated Boolean algebras of 1. equivalence classes and super/sub classes of class algebra type expressions, and 2. a residual Boolean algebra of biclique relationships. It distinguishes which formulas are equivalent, entailed, or unrelated, based on a simplification algorithm that may be thought of as producing a unique pair of Karnaugh maps that describe the rough sets of maximal bicliques of relations. Such maps divide the n-dimensional space of up to 2n-1 conjunctions of up to n propositions into clopen (i.e. a closed set of regions and their boundaries) causal sets. This class algebra is generalized to type-2 fuzzy class algebra by using relative frequencies as probabilities. It is also generalized to a class calculus involving assignments that change the states of programs. INDEX TERMS 4-valued Boolean Logic, Artificial Intelligence, causal sets, class algebra, consciousness, intelligent design, IS-A hierarchy, mathematical logic, meta-theory, pointless topological space, residuated lattices, rough sets, type-2 fuzzy sets |
Tasks | |
Published | 2018-04-10 |
URL | http://arxiv.org/abs/1804.03301v1 |
http://arxiv.org/pdf/1804.03301v1.pdf | |
PWC | https://paperswithcode.com/paper/a-mathematical-framework-for-superintelligent |
Repo | |
Framework | |
Beyond Error Propagation in Neural Machine Translation: Characteristics of Language Also Matter
Title | Beyond Error Propagation in Neural Machine Translation: Characteristics of Language Also Matter |
Authors | Lijun Wu, Xu Tan, Di He, Fei Tian, Tao Qin, Jianhuang Lai, Tie-Yan Liu |
Abstract | Neural machine translation usually adopts autoregressive models and suffers from exposure bias as well as the consequent error propagation problem. Many previous works have discussed the relationship between error propagation and the \emph{accuracy drop} (i.e., the left part of the translated sentence is often better than its right part in left-to-right decoding models) problem. In this paper, we conduct a series of analyses to deeply understand this problem and get several interesting findings. (1) The role of error propagation on accuracy drop is overstated in the literature, although it indeed contributes to the accuracy drop problem. (2) Characteristics of a language play a more important role in causing the accuracy drop: the left part of the translation result in a right-branching language (e.g., English) is more likely to be more accurate than its right part, while the right part is more accurate for a left-branching language (e.g., Japanese). Our discoveries are confirmed on different model structures including Transformer and RNN, and in other sequence generation tasks such as text summarization. |
Tasks | Machine Translation, Text Summarization |
Published | 2018-09-01 |
URL | http://arxiv.org/abs/1809.00120v2 |
http://arxiv.org/pdf/1809.00120v2.pdf | |
PWC | https://paperswithcode.com/paper/beyond-error-propagation-in-neural-machine |
Repo | |
Framework | |
Generalizing Bottleneck Problems
Title | Generalizing Bottleneck Problems |
Authors | Hsiang Hsu, Shahab Asoodeh, Salman Salamatian, Flavio P. Calmon |
Abstract | Given a pair of random variables $(X,Y)\sim P_{XY}$ and two convex functions $f_1$ and $f_2$, we introduce two bottleneck functionals as the lower and upper boundaries of the two-dimensional convex set that consists of the pairs $\left(I_{f_1}(W; X), I_{f_2}(W; Y)\right)$, where $I_f$ denotes $f$-information and $W$ varies over the set of all discrete random variables satisfying the Markov condition $W \to X \to Y$. Applying Witsenhausen and Wyner’s approach, we provide an algorithm for computing boundaries of this set for $f_1$, $f_2$, and discrete $P_{XY}$. In the binary symmetric case, we fully characterize the set when (i) $f_1(t)=f_2(t)=t\log t$, (ii) $f_1(t)=f_2(t)=t^2-1$, and (iii) $f_1$ and $f_2$ are both $\ell^\beta$ norm function for $\beta \geq 2$. We then argue that upper and lower boundaries in (i) correspond to Mrs. Gerber’s Lemma and its inverse (which we call Mr. Gerber’s Lemma), in (ii) correspond to estimation-theoretic variants of Information Bottleneck and Privacy Funnel, and in (iii) correspond to Arimoto Information Bottleneck and Privacy Funnel. |
Tasks | |
Published | 2018-02-16 |
URL | http://arxiv.org/abs/1802.05861v3 |
http://arxiv.org/pdf/1802.05861v3.pdf | |
PWC | https://paperswithcode.com/paper/generalizing-bottleneck-problems |
Repo | |
Framework | |