Paper Group AWR 42
megaman: Manifold Learning with Millions of points. Food Image Recognition by Using Convolutional Neural Networks (CNNs). Neural Semantic Encoders. The Neural Hawkes Process: A Neurally Self-Modulating Multivariate Point Process. The minimal hitting set generation problem: algorithms and computation. A Deep Spatial Contextual Long-term Recurrent Co …
megaman: Manifold Learning with Millions of points
Title | megaman: Manifold Learning with Millions of points |
Authors | James McQueen, Marina Meila, Jacob VanderPlas, Zhongyue Zhang |
Abstract | Manifold Learning is a class of algorithms seeking a low-dimensional non-linear representation of high-dimensional data. Thus manifold learning algorithms are, at least in theory, most applicable to high-dimensional data and sample sizes to enable accurate estimation of the manifold. Despite this, most existing manifold learning implementations are not particularly scalable. Here we present a Python package that implements a variety of manifold learning algorithms in a modular and scalable fashion, using fast approximate neighbors searches and fast sparse eigendecompositions. The package incorporates theoretical advances in manifold learning, such as the unbiased Laplacian estimator and the estimation of the embedding distortion by the Riemannian metric method. In benchmarks, even on a single-core desktop computer, our code embeds millions of data points in minutes, and takes just 200 minutes to embed the main sample of galaxy spectra from the Sloan Digital Sky Survey — consisting of 0.6 million samples in 3750-dimensions — a task which has not previously been possible. |
Tasks | |
Published | 2016-03-09 |
URL | http://arxiv.org/abs/1603.02763v1 |
http://arxiv.org/pdf/1603.02763v1.pdf | |
PWC | https://paperswithcode.com/paper/megaman-manifold-learning-with-millions-of |
Repo | https://github.com/mmp2/megaman |
Framework | none |
Food Image Recognition by Using Convolutional Neural Networks (CNNs)
Title | Food Image Recognition by Using Convolutional Neural Networks (CNNs) |
Authors | Yuzhen Lu |
Abstract | Food image recognition is one of the promising applications of visual object recognition in computer vision. In this study, a small-scale dataset consisting of 5822 images of ten categories and a five-layer CNN was constructed to recognize these images. The bag-of-features (BoF) model coupled with support vector machine (SVM) was first evaluated for image classification, resulting in an overall accuracy of 56%; while the CNN model performed much better with an overall accuracy of 74%. Data augmentation techniques based on geometric transformation were applied to increase the size of training images, which achieved a significantly improved accuracy of more than 90% while preventing the overfitting issue that occurred to the CNN based on raw training data. Further improvements can be expected by collecting more images and optimizing the network architecture and hyper-parameters. |
Tasks | Data Augmentation, Image Classification, Object Recognition |
Published | 2016-12-03 |
URL | http://arxiv.org/abs/1612.00983v2 |
http://arxiv.org/pdf/1612.00983v2.pdf | |
PWC | https://paperswithcode.com/paper/food-image-recognition-by-using-convolutional |
Repo | https://github.com/jingweimo/food-image-classification- |
Framework | none |
Neural Semantic Encoders
Title | Neural Semantic Encoders |
Authors | Tsendsuren Munkhdalai, Hong Yu |
Abstract | We present a memory augmented neural network for natural language understanding: Neural Semantic Encoders. NSE is equipped with a novel memory update rule and has a variable sized encoding memory that evolves over time and maintains the understanding of input sequences through read}, compose and write operations. NSE can also access multiple and shared memories. In this paper, we demonstrated the effectiveness and the flexibility of NSE on five different natural language tasks: natural language inference, question answering, sentence classification, document sentiment analysis and machine translation where NSE achieved state-of-the-art performance when evaluated on publically available benchmarks. For example, our shared-memory model showed an encouraging result on neural machine translation, improving an attention-based baseline by approximately 1.0 BLEU. |
Tasks | Machine Translation, Natural Language Inference, Question Answering, Sentence Classification, Sentiment Analysis |
Published | 2016-07-14 |
URL | http://arxiv.org/abs/1607.04315v3 |
http://arxiv.org/pdf/1607.04315v3.pdf | |
PWC | https://paperswithcode.com/paper/neural-semantic-encoders |
Repo | https://github.com/Smerity/keras_snli |
Framework | none |
The Neural Hawkes Process: A Neurally Self-Modulating Multivariate Point Process
Title | The Neural Hawkes Process: A Neurally Self-Modulating Multivariate Point Process |
Authors | Hongyuan Mei, Jason Eisner |
Abstract | Many events occur in the world. Some event types are stochastically excited or inhibited—in the sense of having their probabilities elevated or decreased—by patterns in the sequence of previous events. Discovering such patterns can help us predict which type of event will happen next and when. We model streams of discrete events in continuous time, by constructing a neurally self-modulating multivariate point process in which the intensities of multiple event types evolve according to a novel continuous-time LSTM. This generative model allows past events to influence the future in complex and realistic ways, by conditioning future event intensities on the hidden state of a recurrent neural network that has consumed the stream of past events. Our model has desirable qualitative properties. It achieves competitive likelihood and predictive accuracy on real and synthetic datasets, including under missing-data conditions. |
Tasks | |
Published | 2016-12-29 |
URL | http://arxiv.org/abs/1612.09328v3 |
http://arxiv.org/pdf/1612.09328v3.pdf | |
PWC | https://paperswithcode.com/paper/the-neural-hawkes-process-a-neurally-self |
Repo | https://github.com/HMEIatJHU/neurawkes |
Framework | none |
The minimal hitting set generation problem: algorithms and computation
Title | The minimal hitting set generation problem: algorithms and computation |
Authors | Andrew Gainer-Dewar, Paola Vera-Licona |
Abstract | Finding inclusion-minimal “hitting sets” for a given collection of sets is a fundamental combinatorial problem with applications in domains as diverse as Boolean algebra, computational biology, and data mining. Much of the algorithmic literature focuses on the problem of recognizing the collection of minimal hitting sets; however, in many of the applications, it is more important to generate these hitting sets. We survey twenty algorithms from across a variety of domains, considering their history, classification, useful features, and computational performance on a variety of synthetic and real-world inputs. We also provide a suite of implementations of these algorithms with a ready-to-use, platform-agnostic interface based on Docker containers and the AlgoRun framework, so that interested computational scientists can easily perform similar tests with inputs from their own research areas on their own computers or through a convenient Web interface. |
Tasks | |
Published | 2016-01-05 |
URL | http://arxiv.org/abs/1601.02939v1 |
http://arxiv.org/pdf/1601.02939v1.pdf | |
PWC | https://paperswithcode.com/paper/the-minimal-hitting-set-generation-problem |
Repo | https://github.com/VeraLiconaResearchGroup/MHSGenerationAlgorithms |
Framework | none |
A Deep Spatial Contextual Long-term Recurrent Convolutional Network for Saliency Detection
Title | A Deep Spatial Contextual Long-term Recurrent Convolutional Network for Saliency Detection |
Authors | Nian Liu, Junwei Han |
Abstract | Traditional saliency models usually adopt hand-crafted image features and human-designed mechanisms to calculate local or global contrast. In this paper, we propose a novel computational saliency model, i.e., deep spatial contextual long-term recurrent convolutional network (DSCLRCN) to predict where people looks in natural scenes. DSCLRCN first automatically learns saliency related local features on each image location in parallel. Then, in contrast with most other deep network based saliency models which infer saliency in local contexts, DSCLRCN can mimic the cortical lateral inhibition mechanisms in human visual system to incorporate global contexts to assess the saliency of each image location by leveraging the deep spatial long short-term memory (DSLSTM) model. Moreover, we also integrate scene context modulation in DSLSTM for saliency inference, leading to a novel deep spatial contextual LSTM (DSCLSTM) model. The whole network can be trained end-to-end and works efficiently when testing. Experimental results on two benchmark datasets show that DSCLRCN can achieve state-of-the-art performance on saliency detection. Furthermore, the proposed DSCLSTM model can significantly boost the saliency detection performance by incorporating both global spatial interconnections and scene context modulation, which may uncover novel inspirations for studies on them in computational saliency models. |
Tasks | Saliency Detection |
Published | 2016-10-06 |
URL | http://arxiv.org/abs/1610.01708v1 |
http://arxiv.org/pdf/1610.01708v1.pdf | |
PWC | https://paperswithcode.com/paper/a-deep-spatial-contextual-long-term-recurrent |
Repo | https://github.com/AAshqar/DSCLRCN-PyTorch |
Framework | pytorch |
Fundamental principles of cortical computation: unsupervised learning with prediction, compression and feedback
Title | Fundamental principles of cortical computation: unsupervised learning with prediction, compression and feedback |
Authors | Micah Richert, Dimitry Fisher, Filip Piekniewski, Eugene M. Izhikevich, Todd L. Hylton |
Abstract | There has been great progress in understanding of anatomical and functional microcircuitry of the primate cortex. However, the fundamental principles of cortical computation - the principles that allow the visual cortex to bind retinal spikes into representations of objects, scenes and scenarios - have so far remained elusive. In an attempt to come closer to understanding the fundamental principles of cortical computation, here we present a functional, phenomenological model of the primate visual cortex. The core part of the model describes four hierarchical cortical areas with feedforward, lateral, and recurrent connections. The three main principles implemented in the model are information compression, unsupervised learning by prediction, and use of lateral and top-down context. We show that the model reproduces key aspects of the primate ventral stream of visual processing including Simple and Complex cells in V1, increasingly complicated feature encoding, and increased separability of object representations in higher cortical areas. The model learns representations of the visual environment that allow for accurate classification and state-of-the-art visual tracking performance on novel objects. |
Tasks | Visual Tracking |
Published | 2016-08-19 |
URL | http://arxiv.org/abs/1608.06277v1 |
http://arxiv.org/pdf/1608.06277v1.pdf | |
PWC | https://paperswithcode.com/paper/fundamental-principles-of-cortical |
Repo | https://github.com/braincorp/ASC |
Framework | none |
De-Conflated Semantic Representations
Title | De-Conflated Semantic Representations |
Authors | Mohammad Taher Pilehvar, Nigel Collier |
Abstract | One major deficiency of most semantic representation techniques is that they usually model a word type as a single point in the semantic space, hence conflating all the meanings that the word can have. Addressing this issue by learning distinct representations for individual meanings of words has been the subject of several research studies in the past few years. However, the generated sense representations are either not linked to any sense inventory or are unreliable for infrequent word senses. We propose a technique that tackles these problems by de-conflating the representations of words based on the deep knowledge it derives from a semantic network. Our approach provides multiple advantages in comparison to the past work, including its high coverage and the ability to generate accurate representations even for infrequent word senses. We carry out evaluations on six datasets across two semantic similarity tasks and report state-of-the-art results on most of them. |
Tasks | Semantic Similarity, Semantic Textual Similarity |
Published | 2016-08-05 |
URL | http://arxiv.org/abs/1608.01961v1 |
http://arxiv.org/pdf/1608.01961v1.pdf | |
PWC | https://paperswithcode.com/paper/de-conflated-semantic-representations |
Repo | https://github.com/pilehvar/deconf |
Framework | none |
Learning long-term dependencies for action recognition with a biologically-inspired deep network
Title | Learning long-term dependencies for action recognition with a biologically-inspired deep network |
Authors | Yemin Shi, Yonghong Tian, Yaowei Wang, Tiejun Huang |
Abstract | Despite a lot of research efforts devoted in recent years, how to efficiently learn long-term dependencies from sequences still remains a pretty challenging task. As one of the key models for sequence learning, recurrent neural network (RNN) and its variants such as long short term memory (LSTM) and gated recurrent unit (GRU) are still not powerful enough in practice. One possible reason is that they have only feedforward connections, which is different from the biological neural system that is typically composed of both feedforward and feedback connections. To address this problem, this paper proposes a biologically-inspired deep network, called shuttleNet\footnote{Our code is available at \url{https://github.com/shiyemin/shuttlenet}}. Technologically, the shuttleNet consists of several processors, each of which is a GRU while associated with multiple groups of cells and states. Unlike traditional RNNs, all processors inside shuttleNet are loop connected to mimic the brain’s feedforward and feedback connections, in which they are shared across multiple pathways in the loop connection. Attention mechanism is then employed to select the best information flow pathway. Extensive experiments conducted on two benchmark datasets (i.e UCF101 and HMDB51) show that we can beat state-of-the-art methods by simply embedding shuttleNet into a CNN-RNN framework. |
Tasks | Temporal Action Localization |
Published | 2016-11-16 |
URL | http://arxiv.org/abs/1611.05216v3 |
http://arxiv.org/pdf/1611.05216v3.pdf | |
PWC | https://paperswithcode.com/paper/learning-long-term-dependencies-for-action |
Repo | https://github.com/shiyemin/shuttlenet |
Framework | tf |
Chinese Poetry Generation with Planning based Neural Network
Title | Chinese Poetry Generation with Planning based Neural Network |
Authors | Zhe Wang, Wei He, Hua Wu, Haiyang Wu, Wei Li, Haifeng Wang, Enhong Chen |
Abstract | Chinese poetry generation is a very challenging task in natural language processing. In this paper, we propose a novel two-stage poetry generating method which first plans the sub-topics of the poem according to the user’s writing intent, and then generates each line of the poem sequentially, using a modified recurrent neural network encoder-decoder framework. The proposed planning-based method can ensure that the generated poem is coherent and semantically consistent with the user’s intent. A comprehensive evaluation with human judgments demonstrates that our proposed approach outperforms the state-of-the-art poetry generating methods and the poem quality is somehow comparable to human poets. |
Tasks | |
Published | 2016-10-31 |
URL | http://arxiv.org/abs/1610.09889v2 |
http://arxiv.org/pdf/1610.09889v2.pdf | |
PWC | https://paperswithcode.com/paper/chinese-poetry-generation-with-planning-based |
Repo | https://github.com/Epoch-Mengying/Generating-Poetry-with-Chatbot |
Framework | none |
Local Binary Convolutional Neural Networks
Title | Local Binary Convolutional Neural Networks |
Authors | Felix Juefei-Xu, Vishnu Naresh Boddeti, Marios Savvides |
Abstract | We propose local binary convolution (LBC), an efficient alternative to convolutional layers in standard convolutional neural networks (CNN). The design principles of LBC are motivated by local binary patterns (LBP). The LBC layer comprises of a set of fixed sparse pre-defined binary convolutional filters that are not updated during the training process, a non-linear activation function and a set of learnable linear weights. The linear weights combine the activated filter responses to approximate the corresponding activated filter responses of a standard convolutional layer. The LBC layer affords significant parameter savings, 9x to 169x in the number of learnable parameters compared to a standard convolutional layer. Furthermore, the sparse and binary nature of the weights also results in up to 9x to 169x savings in model size compared to a standard convolutional layer. We demonstrate both theoretically and experimentally that our local binary convolution layer is a good approximation of a standard convolutional layer. Empirically, CNNs with LBC layers, called local binary convolutional neural networks (LBCNN), achieves performance parity with regular CNNs on a range of visual datasets (MNIST, SVHN, CIFAR-10, and ImageNet) while enjoying significant computational savings. |
Tasks | |
Published | 2016-08-22 |
URL | http://arxiv.org/abs/1608.06049v2 |
http://arxiv.org/pdf/1608.06049v2.pdf | |
PWC | https://paperswithcode.com/paper/local-binary-convolutional-neural-networks |
Repo | https://github.com/juefeix/pnn.pytorch |
Framework | pytorch |
Incorporating Discrete Translation Lexicons into Neural Machine Translation
Title | Incorporating Discrete Translation Lexicons into Neural Machine Translation |
Authors | Philip Arthur, Graham Neubig, Satoshi Nakamura |
Abstract | Neural machine translation (NMT) often makes mistakes in translating low-frequency content words that are essential to understanding the meaning of the sentence. We propose a method to alleviate this problem by augmenting NMT systems with discrete translation lexicons that efficiently encode translations of these low-frequency words. We describe a method to calculate the lexicon probability of the next word in the translation candidate by using the attention vector of the NMT model to select which source word lexical probabilities the model should focus on. We test two methods to combine this probability with the standard NMT probability: (1) using it as a bias, and (2) linear interpolation. Experiments on two corpora show an improvement of 2.0-2.3 BLEU and 0.13-0.44 NIST score, and faster convergence time. |
Tasks | Machine Translation |
Published | 2016-06-07 |
URL | http://arxiv.org/abs/1606.02006v2 |
http://arxiv.org/pdf/1606.02006v2.pdf | |
PWC | https://paperswithcode.com/paper/incorporating-discrete-translation-lexicons |
Repo | https://github.com/duyvuleo/Transformer-DyNet |
Framework | tf |
Learning to Generate Compositional Color Descriptions
Title | Learning to Generate Compositional Color Descriptions |
Authors | Will Monroe, Noah D. Goodman, Christopher Potts |
Abstract | The production of color language is essential for grounded language generation. Color descriptions have many challenging properties: they can be vague, compositionally complex, and denotationally rich. We present an effective approach to generating color descriptions using recurrent neural networks and a Fourier-transformed color representation. Our model outperforms previous work on a conditional language modeling task over a large corpus of naturalistic color descriptions. In addition, probing the model’s output reveals that it can accurately produce not only basic color terms but also descriptors with non-convex denotations (“greenish”), bare modifiers (“bright”, “dull”), and compositional phrases (“faded teal”) not seen in training. |
Tasks | Language Modelling, Text Generation |
Published | 2016-06-13 |
URL | http://arxiv.org/abs/1606.03821v2 |
http://arxiv.org/pdf/1606.03821v2.pdf | |
PWC | https://paperswithcode.com/paper/learning-to-generate-compositional-color |
Repo | https://github.com/stanfordnlp/color-describer |
Framework | none |
Deeper Depth Prediction with Fully Convolutional Residual Networks
Title | Deeper Depth Prediction with Fully Convolutional Residual Networks |
Authors | Iro Laina, Christian Rupprecht, Vasileios Belagiannis, Federico Tombari, Nassir Navab |
Abstract | This paper addresses the problem of estimating the depth map of a scene given a single RGB image. We propose a fully convolutional architecture, encompassing residual learning, to model the ambiguous mapping between monocular images and depth maps. In order to improve the output resolution, we present a novel way to efficiently learn feature map up-sampling within the network. For optimization, we introduce the reverse Huber loss that is particularly suited for the task at hand and driven by the value distributions commonly present in depth maps. Our model is composed of a single architecture that is trained end-to-end and does not rely on post-processing techniques, such as CRFs or other additional refinement steps. As a result, it runs in real-time on images or videos. In the evaluation, we show that the proposed model contains fewer parameters and requires fewer training data than the current state of the art, while outperforming all approaches on depth estimation. Code and models are publicly available. |
Tasks | Depth Estimation |
Published | 2016-06-01 |
URL | http://arxiv.org/abs/1606.00373v2 |
http://arxiv.org/pdf/1606.00373v2.pdf | |
PWC | https://paperswithcode.com/paper/deeper-depth-prediction-with-fully |
Repo | https://github.com/LeonSun0101/CD-SD |
Framework | pytorch |
Learning Convolutional Neural Networks for Graphs
Title | Learning Convolutional Neural Networks for Graphs |
Authors | Mathias Niepert, Mohamed Ahmed, Konstantin Kutzkov |
Abstract | Numerous important problems can be framed as learning from graph data. We propose a framework for learning convolutional neural networks for arbitrary graphs. These graphs may be undirected, directed, and with both discrete and continuous node and edge attributes. Analogous to image-based convolutional networks that operate on locally connected regions of the input, we present a general approach to extracting locally connected regions from graphs. Using established benchmark data sets, we demonstrate that the learned feature representations are competitive with state of the art graph kernels and that their computation is highly efficient. |
Tasks | Graph Classification |
Published | 2016-05-17 |
URL | http://arxiv.org/abs/1605.05273v4 |
http://arxiv.org/pdf/1605.05273v4.pdf | |
PWC | https://paperswithcode.com/paper/learning-convolutional-neural-networks-for |
Repo | https://github.com/TibiG97/Part2Project |
Framework | none |