Paper Group ANR 701
Hybrid Clustering based on Content and Connection Structure using Joint Nonnegative Matrix Factorization. Learning hard quantum distributions with variational autoencoders. A Line-Point Unified Solution to Relative Camera Pose Estimation. The Pragmatics of Indirect Commands in Collaborative Discourse. Markov Decision Processes with Continuous Side …
Hybrid Clustering based on Content and Connection Structure using Joint Nonnegative Matrix Factorization
Title | Hybrid Clustering based on Content and Connection Structure using Joint Nonnegative Matrix Factorization |
Authors | Rundong Du, Barry Drake, Haesun Park |
Abstract | We present a hybrid method for latent information discovery on the data sets containing both text content and connection structure based on constrained low rank approximation. The new method jointly optimizes the Nonnegative Matrix Factorization (NMF) objective function for text clustering and the Symmetric NMF (SymNMF) objective function for graph clustering. We propose an effective algorithm for the joint NMF objective function, based on a block coordinate descent (BCD) framework. The proposed hybrid method discovers content associations via latent connections found using SymNMF. The method can also be applied with a natural conversion of the problem when a hypergraph formulation is used or the content is associated with hypergraph edges. Experimental results show that by simultaneously utilizing both content and connection structure, our hybrid method produces higher quality clustering results compared to the other NMF clustering methods that uses content alone (standard NMF) or connection structure alone (SymNMF). We also present some interesting applications to several types of real world data such as citation recommendations of papers. The hybrid method proposed in this paper can also be applied to general data expressed with both feature space vectors and pairwise similarities and can be extended to the case with multiple feature spaces or multiple similarity measures. |
Tasks | Graph Clustering, Text Clustering |
Published | 2017-03-28 |
URL | http://arxiv.org/abs/1703.09646v1 |
http://arxiv.org/pdf/1703.09646v1.pdf | |
PWC | https://paperswithcode.com/paper/hybrid-clustering-based-on-content-and |
Repo | |
Framework | |
Learning hard quantum distributions with variational autoencoders
Title | Learning hard quantum distributions with variational autoencoders |
Authors | Andrea Rocchetto, Edward Grant, Sergii Strelchuk, Giuseppe Carleo, Simone Severini |
Abstract | Studying general quantum many-body systems is one of the major challenges in modern physics because it requires an amount of computational resources that scales exponentially with the size of the system.Simulating the evolution of a state, or even storing its description, rapidly becomes intractable for exact classical algorithms. Recently, machine learning techniques, in the form of restricted Boltzmann machines, have been proposed as a way to efficiently represent certain quantum states with applications in state tomography and ground state estimation. Here, we introduce a new representation of states based on variational autoencoders. Variational autoencoders are a type of generative model in the form of a neural network. We probe the power of this representation by encoding probability distributions associated with states from different classes. Our simulations show that deep networks give a better representation for states that are hard to sample from, while providing no benefit for random states. This suggests that the probability distributions associated to hard quantum states might have a compositional structure that can be exploited by layered neural networks. Specifically, we consider the learnability of a class of quantum states introduced by Fefferman and Umans. Such states are provably hard to sample for classical computers, but not for quantum ones, under plausible computational complexity assumptions. The good level of compression achieved for hard states suggests these methods can be suitable for characterising states of the size expected in first generation quantum hardware. |
Tasks | |
Published | 2017-10-02 |
URL | http://arxiv.org/abs/1710.00725v2 |
http://arxiv.org/pdf/1710.00725v2.pdf | |
PWC | https://paperswithcode.com/paper/learning-hard-quantum-distributions-with |
Repo | |
Framework | |
A Line-Point Unified Solution to Relative Camera Pose Estimation
Title | A Line-Point Unified Solution to Relative Camera Pose Estimation |
Authors | Ashraf Qadir, Jeremiah Neubert |
Abstract | In this work we present a unified method of relative camera pose estimation from points and lines correspondences. Given a set of 2D points and lines correspondences in three views, of which two are known, a method has been developed for estimating the camera pose of the third view. Novelty of this algorithm is to combine both points and lines correspondences in the camera pose estimation which enables us to compute relative camera pose with a small number of feature correspondences. Our central idea is to exploit the tri-linear relationship between three views and generate a set of linear equations from the points and lines correspondences in the three views. The desired solution to the system of equations are expressed as a linear combination of the singular vectors and the coefficients are computed by solving a small set of quadratic equations generated by imposing orthonormality constraints for general camera motion. The advantages of the proposed method are demonstrated by experimenting on publicly available data set. Results show the robustness and efficiency of the method in relative camera pose estimation for both small and large camera motion with a small set of points and line features. |
Tasks | Pose Estimation |
Published | 2017-10-17 |
URL | http://arxiv.org/abs/1710.06495v1 |
http://arxiv.org/pdf/1710.06495v1.pdf | |
PWC | https://paperswithcode.com/paper/a-line-point-unified-solution-to-relative |
Repo | |
Framework | |
The Pragmatics of Indirect Commands in Collaborative Discourse
Title | The Pragmatics of Indirect Commands in Collaborative Discourse |
Authors | Matthew Lamm, Mihail Eric |
Abstract | Today’s artificial assistants are typically prompted to perform tasks through direct, imperative commands such as \emph{Set a timer} or \emph{Pick up the box}. However, to progress toward more natural exchanges between humans and these assistants, it is important to understand the way non-imperative utterances can indirectly elicit action of an addressee. In this paper, we investigate command types in the setting of a grounded, collaborative game. We focus on a less understood family of utterances for eliciting agent action, locatives like \emph{The chair is in the other room}, and demonstrate how these utterances indirectly command in specific game state contexts. Our work shows that models with domain-specific grounding can effectively realize the pragmatic reasoning that is necessary for more robust natural language interaction. |
Tasks | |
Published | 2017-05-08 |
URL | http://arxiv.org/abs/1705.03454v2 |
http://arxiv.org/pdf/1705.03454v2.pdf | |
PWC | https://paperswithcode.com/paper/the-pragmatics-of-indirect-commands-in |
Repo | |
Framework | |
Markov Decision Processes with Continuous Side Information
Title | Markov Decision Processes with Continuous Side Information |
Authors | Aditya Modi, Nan Jiang, Satinder Singh, Ambuj Tewari |
Abstract | We consider a reinforcement learning (RL) setting in which the agent interacts with a sequence of episodic MDPs. At the start of each episode the agent has access to some side-information or context that determines the dynamics of the MDP for that episode. Our setting is motivated by applications in healthcare where baseline measurements of a patient at the start of a treatment episode form the context that may provide information about how the patient might respond to treatment decisions. We propose algorithms for learning in such Contextual Markov Decision Processes (CMDPs) under an assumption that the unobserved MDP parameters vary smoothly with the observed context. We also give lower and upper PAC bounds under the smoothness assumption. Because our lower bound has an exponential dependence on the dimension, we consider a tractable linear setting where the context is used to create linear combinations of a finite set of MDPs. For the linear setting, we give a PAC learning algorithm based on KWIK learning techniques. |
Tasks | |
Published | 2017-11-15 |
URL | http://arxiv.org/abs/1711.05726v1 |
http://arxiv.org/pdf/1711.05726v1.pdf | |
PWC | https://paperswithcode.com/paper/markov-decision-processes-with-continuous |
Repo | |
Framework | |
Max-Pooling Loss Training of Long Short-Term Memory Networks for Small-Footprint Keyword Spotting
Title | Max-Pooling Loss Training of Long Short-Term Memory Networks for Small-Footprint Keyword Spotting |
Authors | Ming Sun, Anirudh Raju, George Tucker, Sankaran Panchapagesan, Gengshen Fu, Arindam Mandal, Spyros Matsoukas, Nikko Strom, Shiv Vitaladevuni |
Abstract | We propose a max-pooling based loss function for training Long Short-Term Memory (LSTM) networks for small-footprint keyword spotting (KWS), with low CPU, memory, and latency requirements. The max-pooling loss training can be further guided by initializing with a cross-entropy loss trained network. A posterior smoothing based evaluation approach is employed to measure keyword spotting performance. Our experimental results show that LSTM models trained using cross-entropy loss or max-pooling loss outperform a cross-entropy loss trained baseline feed-forward Deep Neural Network (DNN). In addition, max-pooling loss trained LSTM with randomly initialized network performs better compared to cross-entropy loss trained LSTM. Finally, the max-pooling loss trained LSTM initialized with a cross-entropy pre-trained network shows the best performance, which yields $67.6%$ relative reduction compared to baseline feed-forward DNN in Area Under the Curve (AUC) measure. |
Tasks | Keyword Spotting, Small-Footprint Keyword Spotting |
Published | 2017-05-05 |
URL | http://arxiv.org/abs/1705.02411v1 |
http://arxiv.org/pdf/1705.02411v1.pdf | |
PWC | https://paperswithcode.com/paper/max-pooling-loss-training-of-long-short-term |
Repo | |
Framework | |
Conditional Random Field and Deep Feature Learning for Hyperspectral Image Segmentation
Title | Conditional Random Field and Deep Feature Learning for Hyperspectral Image Segmentation |
Authors | Fahim Irfan Alam, Jun Zhou, Alan Wee-Chung Liew, Xiuping Jia, Jocelyn Chanussot, Yongsheng Gao |
Abstract | Image segmentation is considered to be one of the critical tasks in hyperspectral remote sensing image processing. Recently, convolutional neural network (CNN) has established itself as a powerful model in segmentation and classification by demonstrating excellent performances. The use of a graphical model such as a conditional random field (CRF) contributes further in capturing contextual information and thus improving the segmentation performance. In this paper, we propose a method to segment hyperspectral images by considering both spectral and spatial information via a combined framework consisting of CNN and CRF. We use multiple spectral cubes to learn deep features using CNN, and then formulate deep CRF with CNN-based unary and pairwise potential functions to effectively extract the semantic correlations between patches consisting of three-dimensional data cubes. Effective piecewise training is applied in order to avoid the computationally expensive iterative CRF inference. Furthermore, we introduce a deep deconvolution network that improves the segmentation masks. We also introduce a new dataset and experimented our proposed method on it along with several widely adopted benchmark datasets to evaluate the effectiveness of our method. By comparing our results with those from several state-of-the-art models, we show the promising potential of our method. |
Tasks | Hyperspectral Image Segmentation, Semantic Segmentation |
Published | 2017-11-13 |
URL | http://arxiv.org/abs/1711.04483v2 |
http://arxiv.org/pdf/1711.04483v2.pdf | |
PWC | https://paperswithcode.com/paper/conditional-random-field-and-deep-feature |
Repo | |
Framework | |
Memory Enriched Big Bang Big Crunch Optimization Algorithm for Data Clustering
Title | Memory Enriched Big Bang Big Crunch Optimization Algorithm for Data Clustering |
Authors | Kayvan Bijari, Hadi Zare, Hadi Veisi, Hossein Bobarshad |
Abstract | Cluster analysis plays an important role in decision making process for many knowledge-based systems. There exist a wide variety of different approaches for clustering applications including the heuristic techniques, probabilistic models, and traditional hierarchical algorithms. In this paper, a novel heuristic approach based on big bang-big crunch algorithm is proposed for clustering problems. The proposed method not only takes advantage of heuristic nature to alleviate typical clustering algorithms such as k-means, but it also benefits from the memory based scheme as compared to its similar heuristic techniques. Furthermore, the performance of the proposed algorithm is investigated based on several benchmark test functions as well as on the well-known datasets. The experimental results show the significant superiority of the proposed method over the similar algorithms. |
Tasks | Decision Making |
Published | 2017-03-08 |
URL | http://arxiv.org/abs/1703.02883v1 |
http://arxiv.org/pdf/1703.02883v1.pdf | |
PWC | https://paperswithcode.com/paper/memory-enriched-big-bang-big-crunch |
Repo | |
Framework | |
Neural Paraphrase Identification of Questions with Noisy Pretraining
Title | Neural Paraphrase Identification of Questions with Noisy Pretraining |
Authors | Gaurav Singh Tomar, Thyago Duque, Oscar Täckström, Jakob Uszkoreit, Dipanjan Das |
Abstract | We present a solution to the problem of paraphrase identification of questions. We focus on a recent dataset of question pairs annotated with binary paraphrase labels and show that a variant of the decomposable attention model (Parikh et al., 2016) results in accurate performance on this task, while being far simpler than many competing neural architectures. Furthermore, when the model is pretrained on a noisy dataset of automatically collected question paraphrases, it obtains the best reported performance on the dataset. |
Tasks | Paraphrase Identification |
Published | 2017-04-15 |
URL | http://arxiv.org/abs/1704.04565v2 |
http://arxiv.org/pdf/1704.04565v2.pdf | |
PWC | https://paperswithcode.com/paper/neural-paraphrase-identification-of-questions |
Repo | |
Framework | |
A quantum dynamic belief model to explain the interference effects of categorization on decision making
Title | A quantum dynamic belief model to explain the interference effects of categorization on decision making |
Authors | Zichang He, Wen Jiang |
Abstract | Categorization is necessary for many decision making tasks. However, the categorization process may interfere the decision making result and the law of total probability can be violated in some situations. To predict the interference effect of categorization, some model based on quantum probability has been proposed. In this paper, a new quantum dynamic belief (QDB) model is proposed. Considering the precise decision may not be made during the process, the concept of uncertainty is introduced in our model to simulate real human thinking process. Then the interference effect categorization can be predicted by handling the uncertain information. The proposed model is applied to a categorization decision-making experiment to explain the interference effect of categorization. Compared with other models, our model is relatively more succinct and the result shows the correctness and effectiveness of our model. |
Tasks | Decision Making |
Published | 2017-03-06 |
URL | http://arxiv.org/abs/1703.02894v1 |
http://arxiv.org/pdf/1703.02894v1.pdf | |
PWC | https://paperswithcode.com/paper/a-quantum-dynamic-belief-model-to-explain-the |
Repo | |
Framework | |
A simple data discretizer
Title | A simple data discretizer |
Authors | Gourab Mitra, Shashidhar Sundareisan, Bikash Kanti Sarkar |
Abstract | Data discretization is an important step in the process of machine learning, since it is easier for classifiers to deal with discrete attributes rather than continuous attributes. Over the years, several methods of performing discretization such as Boolean Reasoning, Equal Frequency Binning, Entropy have been proposed, explored, and implemented. In this article, a simple supervised discretization approach is introduced. The prime goal of MIL is to maximize classification accuracy of classifier, minimizing loss of information while discretization of continuous attributes. The performance of the suggested approach is compared with the supervised discretization algorithm Minimum Information Loss (MIL), using the state-of-the-art rule inductive algorithms- J48 (Java implementation of C4.5 classifier). The presented approach is, indeed, the modified version of MIL. The empirical results show that the modified approach performs better in several cases in comparison to the original MIL algorithm and Minimum Description Length Principle (MDLP) . |
Tasks | |
Published | 2017-10-13 |
URL | http://arxiv.org/abs/1710.05091v1 |
http://arxiv.org/pdf/1710.05091v1.pdf | |
PWC | https://paperswithcode.com/paper/a-simple-data-discretizer |
Repo | |
Framework | |
Comparison of Different Methods for Tissue Segmentation in Histopathological Whole-Slide Images
Title | Comparison of Different Methods for Tissue Segmentation in Histopathological Whole-Slide Images |
Authors | Péter Bándi, Rob van de Loo, Milad Intezar, Daan Geijs, Francesco Ciompi, Bram van Ginneken, Jeroen van der Laak, Geert Litjens |
Abstract | Tissue segmentation is an important pre-requisite for efficient and accurate diagnostics in digital pathology. However, it is well known that whole-slide scanners can fail in detecting all tissue regions, for example due to the tissue type, or due to weak staining because their tissue detection algorithms are not robust enough. In this paper, we introduce two different convolutional neural network architectures for whole slide image segmentation to accurately identify the tissue sections. We also compare the algorithms to a published traditional method. We collected 54 whole slide images with differing stains and tissue types from three laboratories to validate our algorithms. We show that while the two methods do not differ significantly they outperform their traditional counterpart (Jaccard index of 0.937 and 0.929 vs. 0.870, p < 0.01). |
Tasks | Semantic Segmentation |
Published | 2017-03-17 |
URL | http://arxiv.org/abs/1703.05990v2 |
http://arxiv.org/pdf/1703.05990v2.pdf | |
PWC | https://paperswithcode.com/paper/comparison-of-different-methods-for-tissue |
Repo | |
Framework | |
A multi-agent reinforcement learning model of common-pool resource appropriation
Title | A multi-agent reinforcement learning model of common-pool resource appropriation |
Authors | Julien Perolat, Joel Z. Leibo, Vinicius Zambaldi, Charles Beattie, Karl Tuyls, Thore Graepel |
Abstract | Humanity faces numerous problems of common-pool resource appropriation. This class of multi-agent social dilemma includes the problems of ensuring sustainable use of fresh water, common fisheries, grazing pastures, and irrigation systems. Abstract models of common-pool resource appropriation based on non-cooperative game theory predict that self-interested agents will generally fail to find socially positive equilibria—a phenomenon called the tragedy of the commons. However, in reality, human societies are sometimes able to discover and implement stable cooperative solutions. Decades of behavioral game theory research have sought to uncover aspects of human behavior that make this possible. Most of that work was based on laboratory experiments where participants only make a single choice: how much to appropriate. Recognizing the importance of spatial and temporal resource dynamics, a recent trend has been toward experiments in more complex real-time video game-like environments. However, standard methods of non-cooperative game theory can no longer be used to generate predictions for this case. Here we show that deep reinforcement learning can be used instead. To that end, we study the emergent behavior of groups of independently learning agents in a partially observed Markov game modeling common-pool resource appropriation. Our experiments highlight the importance of trial-and-error learning in common-pool resource appropriation and shed light on the relationship between exclusion, sustainability, and inequality. |
Tasks | Multi-agent Reinforcement Learning |
Published | 2017-07-20 |
URL | http://arxiv.org/abs/1707.06600v2 |
http://arxiv.org/pdf/1707.06600v2.pdf | |
PWC | https://paperswithcode.com/paper/a-multi-agent-reinforcement-learning-model-of |
Repo | |
Framework | |
Improving Implicit Semantic Role Labeling by Predicting Semantic Frame Arguments
Title | Improving Implicit Semantic Role Labeling by Predicting Semantic Frame Arguments |
Authors | Quynh Ngoc Thi Do, Steven Bethard, Marie-Francine Moens |
Abstract | Implicit semantic role labeling (iSRL) is the task of predicting the semantic roles of a predicate that do not appear as explicit arguments, but rather regard common sense knowledge or are mentioned earlier in the discourse. We introduce an approach to iSRL based on a predictive recurrent neural semantic frame model (PRNSFM) that uses a large unannotated corpus to learn the probability of a sequence of semantic arguments given a predicate. We leverage the sequence probabilities predicted by the PRNSFM to estimate selectional preferences for predicates and their arguments. On the NomBank iSRL test set, our approach improves state-of-the-art performance on implicit semantic role labeling with less reliance than prior work on manually constructed language resources. |
Tasks | Common Sense Reasoning, Semantic Role Labeling |
Published | 2017-04-10 |
URL | http://arxiv.org/abs/1704.02709v2 |
http://arxiv.org/pdf/1704.02709v2.pdf | |
PWC | https://paperswithcode.com/paper/improving-implicit-semantic-role-labeling-by |
Repo | |
Framework | |
GeoSeq2Seq: Information Geometric Sequence-to-Sequence Networks
Title | GeoSeq2Seq: Information Geometric Sequence-to-Sequence Networks |
Authors | Alessandro Bay, Biswa Sengupta |
Abstract | The Fisher information metric is an important foundation of information geometry, wherein it allows us to approximate the local geometry of a probability distribution. Recurrent neural networks such as the Sequence-to-Sequence (Seq2Seq) networks that have lately been used to yield state-of-the-art performance on speech translation or image captioning have so far ignored the geometry of the latent embedding, that they iteratively learn. We propose the information geometric Seq2Seq (GeoSeq2Seq) network which abridges the gap between deep recurrent neural networks and information geometry. Specifically, the latent embedding offered by a recurrent network is encoded as a Fisher kernel of a parametric Gaussian Mixture Model, a formalism common in computer vision. We utilise such a network to predict the shortest routes between two nodes of a graph by learning the adjacency matrix using the GeoSeq2Seq formalism; our results show that for such a problem the probabilistic representation of the latent embedding supersedes the non-probabilistic embedding by 10-15%. |
Tasks | Image Captioning |
Published | 2017-10-25 |
URL | http://arxiv.org/abs/1710.09363v2 |
http://arxiv.org/pdf/1710.09363v2.pdf | |
PWC | https://paperswithcode.com/paper/geoseq2seq-information-geometric-sequence-to |
Repo | |
Framework | |