Paper Group ANR 171
An Axiomatic Approach to Regularizing Neural Ranking Models. Dense RepPoints: Representing Visual Objects with Dense Point Sets. Opportunities for artificial intelligence in advancing precision medicine. A Neurally-Inspired Hierarchical Prediction Network for Spatiotemporal Sequence Learning and Prediction. Neural Review Rating Prediction with Hier …
An Axiomatic Approach to Regularizing Neural Ranking Models
Title | An Axiomatic Approach to Regularizing Neural Ranking Models |
Authors | Corby Rosset, Bhaskar Mitra, Chenyan Xiong, Nick Craswell, Xia Song, Saurabh Tiwary |
Abstract | Axiomatic information retrieval (IR) seeks a set of principle properties desirable in IR models. These properties when formally expressed provide guidance in the search for better relevance estimation functions. Neural ranking models typically contain a large number of parameters. The training of these models involve a search for appropriate parameter values based on large quantities of labeled examples. Intuitively, axioms that can guide the search for better traditional IR models should also help in better parameter estimation for machine learning based rankers. This work explores the use of IR axioms to augment the direct supervision from labeled data for training neural ranking models. We modify the documents in our dataset along the lines of well-known axioms during training and add a regularization loss based on the agreement between the ranking model and the axioms on which version of the document—the original or the perturbed—should be preferred. Our experiments show that the neural ranking model achieves faster convergence and better generalization with axiomatic regularization. |
Tasks | Information Retrieval |
Published | 2019-04-15 |
URL | http://arxiv.org/abs/1904.06808v1 |
http://arxiv.org/pdf/1904.06808v1.pdf | |
PWC | https://paperswithcode.com/paper/an-axiomatic-approach-to-regularizing-neural |
Repo | |
Framework | |
Dense RepPoints: Representing Visual Objects with Dense Point Sets
Title | Dense RepPoints: Representing Visual Objects with Dense Point Sets |
Authors | Ze Yang, Yinghao Xu, Han Xue, Zheng Zhang, Raquel Urtasun, Liwei Wang, Stephen Lin, Han Hu |
Abstract | We present an object representation, called \textbf{Dense RepPoints}, for flexible and detailed modeling of object appearance and geometry. In contrast to the coarse geometric localization and feature extraction of bounding boxes, Dense RepPoints adaptively distributes a dense set of points to semantically and geometrically significant positions on an object, providing informative cues for object analysis. Techniques are developed to address challenges related to supervised training for dense point sets from image segments annotations and making this extensive representation computationally practical. In addition, the versatility of this representation is exploited to model object structure over multiple levels of granularity. Dense RepPoints significantly improves performance on geometrically-oriented visual understanding tasks, including a $1.6$ AP gain in object detection on the challenging COCO benchmark. |
Tasks | Object Detection |
Published | 2019-12-24 |
URL | https://arxiv.org/abs/1912.11473v1 |
https://arxiv.org/pdf/1912.11473v1.pdf | |
PWC | https://paperswithcode.com/paper/dense-reppoints-representing-visual-objects |
Repo | |
Framework | |
Opportunities for artificial intelligence in advancing precision medicine
Title | Opportunities for artificial intelligence in advancing precision medicine |
Authors | Fabian V. Filipp |
Abstract | Machine learning (ML), deep learning (DL), and artificial intelligence (AI) are of increasing importance in biomedicine. The goal of this work is to show progress in ML in digital health, to exemplify future needs and trends, and to identify any essential prerequisites of AI and ML for precision health. High-throughput technologies are delivering growing volumes of biomedical data, such as large-scale genome-wide sequencing assays, libraries of medical images, or drug perturbation screens of healthy, developing, and diseased tissue. Multi-omics data in biomedicine is deep and complex, offering an opportunity for data-driven insights and automated disease classification. Learning from these data will open our understanding and definition of healthy baselines and disease signatures. State-of-the-art applications of deep neural networks include digital image recognition, single cell clustering, and virtual drug screens, demonstrating breadths and power of ML in biomedicine. Significantly, AI and systems biology have embraced big data challenges and may enable novel biotechnology-derived therapies to facilitate the implementation of precision medicine approaches. |
Tasks | |
Published | 2019-11-17 |
URL | https://arxiv.org/abs/1911.07125v1 |
https://arxiv.org/pdf/1911.07125v1.pdf | |
PWC | https://paperswithcode.com/paper/opportunities-for-artificial-intelligence-in |
Repo | |
Framework | |
A Neurally-Inspired Hierarchical Prediction Network for Spatiotemporal Sequence Learning and Prediction
Title | A Neurally-Inspired Hierarchical Prediction Network for Spatiotemporal Sequence Learning and Prediction |
Authors | Jielin Qiu, Ge Huang, Tai Sing Lee |
Abstract | In this paper we developed a hierarchical network model, called Hierarchical Prediction Network (HPNet), to understand how spatiotemporal memories might be learned and encoded in the recurrent circuits in the visual cortical hierarchy for predicting future video frames. This neurally inspired model operates in the analysis-by-synthesis framework. It contains a feed-forward path that computes and encodes spatiotemporal features of successive complexity and a feedback path for the successive levels to project their interpretations to the level below. Within each level, the feed-forward path and the feedback path intersect in a recurrent gated circuit, instantiated in a LSTM module, to generate a prediction or explanation of the incoming signals. The network learns its internal model of the world by minimizing the errors of its prediction of the incoming signals at each level of the hierarchy. We found that hierarchical interaction in the network increases semantic clustering of global movement patterns in the population codes of the units along the hierarchy, even in the earliest module. This facilitates the learning of relationships among movement patterns, yielding state-of-the-art performance in long range video sequence predictions in the benchmark datasets. The network model automatically reproduces a variety of prediction suppression and familiarity suppression neurophysiological phenomena observed in the visual cortex, suggesting that hierarchical prediction might indeed be an important principle for representational learning in the visual cortex. |
Tasks | |
Published | 2019-01-25 |
URL | http://arxiv.org/abs/1901.09002v1 |
http://arxiv.org/pdf/1901.09002v1.pdf | |
PWC | https://paperswithcode.com/paper/a-neurally-inspired-hierarchical-prediction |
Repo | |
Framework | |
Neural Review Rating Prediction with Hierarchical Attentions and Latent Factors
Title | Neural Review Rating Prediction with Hierarchical Attentions and Latent Factors |
Authors | Xianchen Wang, Hongtao Liu, Peiyi Wang, Fangzhao Wu, Hongyan Xu, Wenjun Wang, Xing Xie |
Abstract | Text reviews can provide rich useful semantic information for modeling users and items, which can benefit rating prediction in recommendation. Different words and reviews may have different informativeness for users or items. Besides, different users and items should be personalized. Most existing works regard all reviews equally or utilize a general attention mechanism. In this paper, we propose a hierarchical attention model fusing latent factor model for rating prediction with reviews, which can focus on important words and informative reviews. Specially, we use the factor vectors of Latent Factor Model to guide the attention network and combine the factor vectors with feature representation learned from reviews to predict the final ratings. Experiments on real-world datasets validate the effectiveness of our approach. |
Tasks | |
Published | 2019-05-29 |
URL | https://arxiv.org/abs/1906.01511v1 |
https://arxiv.org/pdf/1906.01511v1.pdf | |
PWC | https://paperswithcode.com/paper/190601511 |
Repo | |
Framework | |
Collaboration based Multi-Label Learning
Title | Collaboration based Multi-Label Learning |
Authors | Lei Feng, Bo An, Shuo He |
Abstract | It is well-known that exploiting label correlations is crucially important to multi-label learning. Most of the existing approaches take label correlations as prior knowledge, which may not correctly characterize the real relationships among labels. Besides, label correlations are normally used to regularize the hypothesis space, while the final predictions are not explicitly correlated. In this paper, we suggest that for each individual label, the final prediction involves the collaboration between its own prediction and the predictions of other labels. Based on this assumption, we first propose a novel method to learn the label correlations via sparse reconstruction in the label space. Then, by seamlessly integrating the learned label correlations into model training, we propose a novel multi-label learning approach that aims to explicitly account for the correlated predictions of labels while training the desired model simultaneously. Extensive experimental results show that our approach outperforms the state-of-the-art counterparts. |
Tasks | Multi-Label Learning |
Published | 2019-02-08 |
URL | http://arxiv.org/abs/1902.03047v1 |
http://arxiv.org/pdf/1902.03047v1.pdf | |
PWC | https://paperswithcode.com/paper/collaboration-based-multi-label-learning |
Repo | |
Framework | |
Value-Added Chemical Discovery Using Reinforcement Learning
Title | Value-Added Chemical Discovery Using Reinforcement Learning |
Authors | Peihong Jiang, Hieu Doan, Sandeep Madireddy, Rajeev Surendran Assary, Prasanna Balaprakash |
Abstract | Computer-assisted synthesis planning aims to help chemists find better reaction pathways faster. Finding viable and short pathways from sugar molecules to value-added chemicals can be modeled as a retrosynthesis planning problem with a catalyst allowed. This is a crucial step in efficient biomass conversion. The traditional computational chemistry approach to identifying possible reaction pathways involves computing the reaction energies of hundreds of intermediates, which is a critical bottleneck in silico reaction discovery. Deep reinforcement learning has shown in other domains that a well-trained agent with little or no prior human knowledge can surpass human performance. While some effort has been made to adapt machine learning techniques to the retrosynthesis planning problem, value-added chemical discovery presents unique challenges. Specifically, the reaction can occur in several different sites in a molecule, a subtle case that has never been treated in previous works. With a more versatile formulation of the problem as a Markov decision process, we address the problem using deep reinforcement learning techniques and present promising preliminary results. |
Tasks | |
Published | 2019-11-10 |
URL | https://arxiv.org/abs/1911.07630v1 |
https://arxiv.org/pdf/1911.07630v1.pdf | |
PWC | https://paperswithcode.com/paper/value-added-chemical-discovery-using |
Repo | |
Framework | |
Recursive Prediction of Graph Signals with Incoming Nodes
Title | Recursive Prediction of Graph Signals with Incoming Nodes |
Authors | Arun Venkitaraman, Saikat Chatterjee, Bo Wahlberg |
Abstract | Kernel and linear regression have been recently explored in the prediction of graph signals as the output, given arbitrary input signals that are agnostic to the graph. In many real-world problems, the graph expands over time as new nodes get introduced. Keeping this premise in mind, we propose a method to recursively obtain the optimal prediction or regression coefficients for the recently propose Linear Regression over Graphs (LRG), as the graph expands with incoming nodes. This comes as a natural consequence of the structure C(W)= of the regression problem, and obviates the need to solve a new regression problem each time a new node is added. Experiments with real-world graph signals show that our approach results in good prediction performance which tends to be close to that obtained from knowing the entire graph apriori. |
Tasks | |
Published | 2019-11-26 |
URL | https://arxiv.org/abs/1911.11542v1 |
https://arxiv.org/pdf/1911.11542v1.pdf | |
PWC | https://paperswithcode.com/paper/recursive-prediction-of-graph-signals-with |
Repo | |
Framework | |
A multi-path 2.5 dimensional convolutional neural network system for segmenting stroke lesions in brain MRI images
Title | A multi-path 2.5 dimensional convolutional neural network system for segmenting stroke lesions in brain MRI images |
Authors | Yunzhe Xue, Fadi G. Farhat, Olga Boukrina, A . M. Barrett, Jeffrey R. Binder, Usman W. Roshan, William W. Graves |
Abstract | Automatic identification of brain lesions from magnetic resonance imaging (MRI) scans of stroke survivors would be a useful aid in patient diagnosis and treatment planning. We propose a multi-modal multi-path convolutional neural network system for automating stroke lesion segmentation. Our system has nine end-to-end UNets that take as input 2-dimensional (2D) slices and examines all three planes with three different normalizations. Outputs from these nine total paths are concatenated into a 3D volume that is then passed to a 3D convolutional neural network to output a final lesion mask. We trained and tested our method on datasets from three sources: Medical College of Wisconsin (MCW), Kessler Foundation (KF), and the publicly available Anatomical Tracings of Lesions After Stroke (ATLAS) dataset. Cross-study validation results (with independent training and validation datasets) were obtained to compare with previous methods based on naive Bayes, random forests, and three recently published convolutional neural networks. Model performance was quantified in terms of the Dice coefficient. Training on the KF and MCW images and testing on the ATLAS images yielded a mean Dice coefficient of 0.54. This was reliably better than the next best previous model, UNet, at 0.47. Reversing the train and test datasets yields a mean Dice of 0.47 on KF and MCW images, whereas the next best UNet reaches 0.45. With all three datasets combined, the current system compared to previous methods also attained a reliably higher cross-validation accuracy. It also achieved high Dice values for many smaller lesions that existing methods have difficulty identifying. Overall, our system is a clear improvement over previous methods for automating stroke lesion segmentation, bringing us an important step closer to the inter-rater accuracy level of human experts. |
Tasks | Lesion Segmentation |
Published | 2019-05-26 |
URL | https://arxiv.org/abs/1905.10835v1 |
https://arxiv.org/pdf/1905.10835v1.pdf | |
PWC | https://paperswithcode.com/paper/a-multi-path-25-dimensional-convolutional |
Repo | |
Framework | |
Learning Factored Markov Decision Processes with Unawareness
Title | Learning Factored Markov Decision Processes with Unawareness |
Authors | Craig Innes, Alex Lascarides |
Abstract | Methods for learning and planning in sequential decision problems often assume the learner is aware of all possible states and actions in advance. This assumption is sometimes untenable. In this paper, we give a method to learn factored markov decision problems from both domain exploration and expert assistance, which guarantees convergence to near-optimal behaviour, even when the agent begins unaware of factors critical to success. Our experiments show our agent learns optimal behaviour on small and large problems, and that conserving information on discovering new possibilities results in faster convergence. |
Tasks | |
Published | 2019-02-27 |
URL | http://arxiv.org/abs/1902.10619v1 |
http://arxiv.org/pdf/1902.10619v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-factored-markov-decision-processes |
Repo | |
Framework | |
Gmail Smart Compose: Real-Time Assisted Writing
Title | Gmail Smart Compose: Real-Time Assisted Writing |
Authors | Mia Xu Chen, Benjamin N Lee, Gagan Bansal, Yuan Cao, Shuyuan Zhang, Justin Lu, Jackie Tsay, Yinan Wang, Andrew M. Dai, Zhifeng Chen, Timothy Sohn, Yonghui Wu |
Abstract | In this paper, we present Smart Compose, a novel system for generating interactive, real-time suggestions in Gmail that assists users in writing mails by reducing repetitive typing. In the design and deployment of such a large-scale and complicated system, we faced several challenges including model selection, performance evaluation, serving and other practical issues. At the core of Smart Compose is a large-scale neural language model. We leveraged state-of-the-art machine learning techniques for language model training which enabled high-quality suggestion prediction, and constructed novel serving infrastructure for high-throughput and real-time inference. Experimental results show the effectiveness of our proposed system design and deployment approach. This system is currently being served in Gmail. |
Tasks | Language Modelling, Model Selection |
Published | 2019-05-17 |
URL | https://arxiv.org/abs/1906.00080v1 |
https://arxiv.org/pdf/1906.00080v1.pdf | |
PWC | https://paperswithcode.com/paper/190600080 |
Repo | |
Framework | |
On-chip Few-shot Learning with Surrogate Gradient Descent on a Neuromorphic Processor
Title | On-chip Few-shot Learning with Surrogate Gradient Descent on a Neuromorphic Processor |
Authors | Kenneth Stewart, Garrick Orchard, Sumit Bam Shrestha, Emre Neftci |
Abstract | Recent work suggests that synaptic plasticity dynamics in biological models of neurons and neuromorphic hardware are compatible with gradient-based learning (Neftci et al., 2019). Gradient-based learning requires iterating several times over a dataset, which is both time-consuming and constrains the training samples to be independently and identically distributed. This is incompatible with learning systems that do not have boundaries between training and inference, such as in neuromorphic hardware. One approach to overcome these constraints is transfer learning, where a portion of the network is pre-trained and mapped into hardware and the remaining portion is trained online. Transfer learning has the advantage that pre-training can be accelerated offline if the task domain is known, and few samples of each class are sufficient for learning the target task at reasonable accuracies. Here, we demonstrate on-line surrogate gradient few-shot learning on Intel’s Loihi neuromorphic research processor using features pre-trained with spike-based gradient backpropagation-through-time. Our experimental results show that the Loihi chip can learn gestures online using a small number of shots and achieve results that are comparable to the models simulated on a conventional processor. |
Tasks | Few-Shot Learning, Transfer Learning |
Published | 2019-10-11 |
URL | https://arxiv.org/abs/1910.04972v6 |
https://arxiv.org/pdf/1910.04972v6.pdf | |
PWC | https://paperswithcode.com/paper/on-chip-few-shot-learning-with-surrogate |
Repo | |
Framework | |
Spatially Regularized Parametric Map Reconstruction for Fast Magnetic Resonance Fingerprinting
Title | Spatially Regularized Parametric Map Reconstruction for Fast Magnetic Resonance Fingerprinting |
Authors | Fabian Balsiger, Alain Jungo, Olivier Scheidegger, Pierre G. Carlier, Mauricio Reyes, Benjamin Marty |
Abstract | Magnetic resonance fingerprinting (MRF) provides a unique concept for simultaneous and fast acquisition of multiple quantitative MR parameters. Despite acquisition efficiency, adoption of MRF into the clinics is hindered by its dictionary-based reconstruction, which is computationally demanding and lacks scalability. Here, we propose a convolutional neural network-based reconstruction, which enables both accurate and fast reconstruction of parametric maps, and is adaptable based on the needs of spatial regularization and the capacity for the reconstruction. We evaluated the method using MRF T1-FF, an MRF sequence for T1 relaxation time of water and fat fraction mapping. We demonstrate the method’s performance on a highly heterogeneous dataset consisting of 164 patients with various neuromuscular diseases imaged at thighs and legs. We empirically show the benefit of incorporating spatial regularization during the reconstruction and demonstrate that the method learns meaningful features from MR physics perspective. Further, we investigate the ability of the method to handle highly heterogeneous morphometric variations and its generalization to anatomical regions unseen during training. The obtained results outperform the state-of-the-art in deep learning-based MRF reconstruction. Coupled with fast MRF sequences, the proposed method has the potential of enabling multiparametric MR imaging in clinically feasible time. |
Tasks | Magnetic Resonance Fingerprinting |
Published | 2019-11-09 |
URL | https://arxiv.org/abs/1911.03786v1 |
https://arxiv.org/pdf/1911.03786v1.pdf | |
PWC | https://paperswithcode.com/paper/spatially-regularized-parametric-map |
Repo | |
Framework | |
Ocular Diseases Diagnosis in Fundus Images using a Deep Learning: Approaches, tools and Performance evaluation
Title | Ocular Diseases Diagnosis in Fundus Images using a Deep Learning: Approaches, tools and Performance evaluation |
Authors | Yaroub Elloumi, Mohamed Akil, Henda Boudegga |
Abstract | Ocular pathology detection from fundus images presents an important challenge on health care. In fact, each pathology has different severity stages that may be deduced by verifying the existence of specific lesions. Each lesion is characterized by morphological features. Moreover, several lesions of different pathologies have similar features. We note that patient may be affected simultaneously by several pathologies. Consequently, the ocular pathology detection presents a multi-class classification with a complex resolution principle. Several detection methods of ocular pathologies from fundus images have been proposed. The methods based on deep learning are distinguished by higher performance detection, due to their capability to configure the network with respect to the detection objective. This work proposes a survey of ocular pathology detection methods based on deep learning. First, we study the existing methods either for lesion segmentation or pathology classification. Afterwards, we extract the principle steps of processing and we analyze the proposed neural network structures. Subsequently, we identify the hardware and software environment required to employ the deep learning architecture. Thereafter, we investigate about the experimentation principles involved to evaluate the methods and the databases used either for training and testing phases. The detection performance ratios and execution times are also reported and discussed. |
Tasks | Lesion Segmentation |
Published | 2019-05-07 |
URL | https://arxiv.org/abs/1905.02544v1 |
https://arxiv.org/pdf/1905.02544v1.pdf | |
PWC | https://paperswithcode.com/paper/ocular-diseases-diagnosis-in-fundus-images |
Repo | |
Framework | |
SimVAE: Simulator-Assisted Training forInterpretable Generative Models
Title | SimVAE: Simulator-Assisted Training forInterpretable Generative Models |
Authors | Akash Srivastava, Jessie Rosenberg, Dan Gutfreund, David D. Cox |
Abstract | This paper presents a simulator-assisted training method (SimVAE) for variational autoencoders (VAE) that leads to a disentangled and interpretable latent space. Training SimVAE is a two-step process in which first a deep generator network(decoder) is trained to approximate the simulator. During this step, the simulator acts as the data source or as a teacher network. Then an inference network (encoder)is trained to invert the decoder. As such, upon complete training, the encoder represents an approximately inverted simulator. By decoupling the training of the encoder and decoder we bypass some of the difficulties that arise in training generative models such as VAEs and generative adversarial networks (GANs). We show applications of our approach in a variety of domains such as circuit design, graphics de-rendering and other natural science problems that involve inference via simulation. |
Tasks | |
Published | 2019-11-19 |
URL | https://arxiv.org/abs/1911.08051v1 |
https://arxiv.org/pdf/1911.08051v1.pdf | |
PWC | https://paperswithcode.com/paper/simvae-simulator-assisted-training |
Repo | |
Framework | |