January 27, 2020

2907 words 14 mins read

Paper Group ANR 1224

Paper Group ANR 1224

Machine learning for protein folding and dynamics. Reliable Weakly Supervised Learning: Maximize Gain and Maintain Safeness. The Power of Factorization Mechanisms in Local and Central Differential Privacy. Neural Machine Translation with Noisy Lexical Constraints. Meta-neural-network for Realtime and Passive Deep-learning-based Object Recognition. …

Machine learning for protein folding and dynamics

Title Machine learning for protein folding and dynamics
Authors Frank Noé, Gianni De Fabritiis, Cecilia Clementi
Abstract Many aspects of the study of protein folding and dynamics have been affected by the recent advances in machine learning. Methods for the prediction of protein structures from their sequences are now heavily based on machine learning tools. The way simulations are performed to explore the energy landscape of protein systems is also changing as force-fields are started to be designed by means of machine learning methods. These methods are also used to extract the essential information from large simulation datasets and to enhance the sampling of rare events such as folding/unfolding transitions. While significant challenges still need to be tackled, we expect these methods to play an important role on the study of protein folding and dynamics in the near future. We discuss here the recent advances on all these fronts and the questions that need to be addressed for machine learning approaches to become mainstream in protein simulation.
Tasks
Published 2019-11-22
URL https://arxiv.org/abs/1911.09811v1
PDF https://arxiv.org/pdf/1911.09811v1.pdf
PWC https://paperswithcode.com/paper/machine-learning-for-protein-folding-and
Repo
Framework

Reliable Weakly Supervised Learning: Maximize Gain and Maintain Safeness

Title Reliable Weakly Supervised Learning: Maximize Gain and Maintain Safeness
Authors Lan-Zhe Guo, Yu-Feng Li, Ming Li, Jin-Feng Yi, Bo-Wen Zhou, Zhi-Hua Zhou
Abstract Weakly supervised data are widespread and have attracted much attention. However, since label quality is often difficult to guarantee, sometimes the use of weakly supervised data will lead to unsatisfactory performance, i.e., performance degradation or poor performance gains. Moreover, it is usually not feasible to manually increase the label quality, which results in weakly supervised learning being somewhat difficult to rely on. In view of this crucial issue, this paper proposes a simple and novel weakly supervised learning framework. We guide the optimization of label quality through a small amount of validation data, and to ensure the safeness of performance while maximizing performance gain. As validation set is a good approximation for describing generalization risk, it can effectively avoid the unsatisfactory performance caused by incorrect data distribution assumptions. We formalize this underlying consideration into a novel Bi-Level optimization and give an effective solution. Extensive experimental results verify that the new framework achieves impressive performance on weakly supervised learning with a small amount of validation data.
Tasks
Published 2019-04-22
URL http://arxiv.org/abs/1904.09743v1
PDF http://arxiv.org/pdf/1904.09743v1.pdf
PWC https://paperswithcode.com/paper/reliable-weakly-supervised-learning-maximize
Repo
Framework

The Power of Factorization Mechanisms in Local and Central Differential Privacy

Title The Power of Factorization Mechanisms in Local and Central Differential Privacy
Authors Alexander Edmonds, Aleksandar Nikolov, Jonathan Ullman
Abstract We give new characterizations of the sample complexity of answering linear queries (statistical queries) in the local and central models of differential privacy: *In the non-interactive local model, we give the first approximate characterization of the sample complexity. Informally our bounds are tight to within polylogarithmic factors in the number of queries and desired accuracy. Our characterization extends to agnostic learning in the local model. *In the central model, we give a characterization of the sample complexity in the high-accuracy regime that is analogous to that of Nikolov, Talwar, and Zhang (STOC 2013), but is both quantitatively tighter and has a dramatically simpler proof. Our lower bounds apply equally to the empirical and population estimation problems. In both cases, our characterizations show that a particular factorization mechanism is approximately optimal, and the optimal sample complexity is bounded from above and below by well studied factorization norms of a matrix associated with the queries.
Tasks
Published 2019-11-19
URL https://arxiv.org/abs/1911.08339v1
PDF https://arxiv.org/pdf/1911.08339v1.pdf
PWC https://paperswithcode.com/paper/the-power-of-factorization-mechanisms-in
Repo
Framework

Neural Machine Translation with Noisy Lexical Constraints

Title Neural Machine Translation with Noisy Lexical Constraints
Authors Huayang Li, Guoping Huang, Lemao Liu
Abstract Lexically constrained decoding for machine translation has shown to be beneficial in previous studies. Unfortunately, constraints provided by users may contain mistakes in real-world situations. It is still an open question that how to manipulate these noisy constraints in such practical scenarios. We present a novel framework that treats constraints as external memories. In this soft manner, a mistaken constraint can be corrected. Experiments demonstrate that our approach can achieve substantial BLEU gains in handling noisy constraints. These results motivate us to apply the proposed approach on a new scenario where constraints are generated without the help of users. Experiments show that our approach can indeed improve the translation quality with the automatically generated constraints.
Tasks Machine Translation
Published 2019-08-13
URL https://arxiv.org/abs/1908.04664v1
PDF https://arxiv.org/pdf/1908.04664v1.pdf
PWC https://paperswithcode.com/paper/neural-machine-translation-with-noisy-lexical
Repo
Framework

Meta-neural-network for Realtime and Passive Deep-learning-based Object Recognition

Title Meta-neural-network for Realtime and Passive Deep-learning-based Object Recognition
Authors Jingkai Weng, Yujiang Ding, Chengbo Hu, Xue-feng Zhu, Bin Liang, Jing Yang, Jianchun Cheng
Abstract Deep-learning recently show great success across disciplines yet conventionally require time-consuming computer processing or bulky-sized diffractive elements. Here we theoretically propose and experimentally demonstrate a purely-passive “meta-neural-network” with compactness and high-resolution for real-time recognizing complicated objects by analyzing acoustic scattering. We prove our meta-neural-network mimics standard neural network despite its small footprint, thanks to unique capability of its metamaterial unit cells, dubbed “meta-neurons”, to produce deep-subwavelength-distribution of discrete phase shift as learnable parameters during training. The resulting device exhibits the “intelligence” to perform desired tasks with potential to address the current trade-off between reducing device’s size, cost and energy consumption and increasing recognition speed and accuracy, showcased by an example of handwritten digit recognition. Our mechanism opens the route to new metamaterial-based deep-learning paradigms and enable conceptual devices such as smart transducers automatically analyzing signals, with far-reaching implications for acoustics, optics and related fields.
Tasks Handwritten Digit Recognition, Object Recognition
Published 2019-09-16
URL https://arxiv.org/abs/1909.07122v1
PDF https://arxiv.org/pdf/1909.07122v1.pdf
PWC https://paperswithcode.com/paper/meta-neural-network-for-realtime-and-passive
Repo
Framework

A Generate-Validate Approach to Answering Questions about Qualitative Relationships

Title A Generate-Validate Approach to Answering Questions about Qualitative Relationships
Authors Arindam Mitra, Chitta Baral, Aurgho Bhattacharjee, Ishan Shrivastava
Abstract Qualitative relationships describe how increasing or decreasing one property (e.g. altitude) affects another (e.g. temperature). They are an important aspect of natural language question answering and are crucial for building chatbots or voice agents where one may enquire about qualitative relationships. Recently a dataset about question answering involving qualitative relationships has been proposed, and a few approaches to answer such questions have been explored, in the heart of which lies a semantic parser that converts the natural language input to a suitable logical form. A problem with existing semantic parsers is that they try to directly convert the input sentences to a logical form. Since the output language varies with each application, it forces the semantic parser to learn almost everything from scratch. In this paper, we show that instead of using a semantic parser to produce the logical form, if we apply the generate-validate framework i.e. generate a natural language description of the logical form and validate if the natural language description is followed from the input text, we get a better scope for transfer learning and our method outperforms the state-of-the-art by a large margin of 7.93%.
Tasks Question Answering, Transfer Learning
Published 2019-08-09
URL https://arxiv.org/abs/1908.03645v1
PDF https://arxiv.org/pdf/1908.03645v1.pdf
PWC https://paperswithcode.com/paper/a-generate-validate-approach-to-answering
Repo
Framework

TEQUILA: Temporal Question Answering over Knowledge Bases

Title TEQUILA: Temporal Question Answering over Knowledge Bases
Authors Zhen Jia, Abdalghani Abujabal, Rishiraj Saha Roy, Jannik Stroetgen, Gerhard Weikum
Abstract Question answering over knowledge bases (KB-QA) poses challenges in handling complex questions that need to be decomposed into sub-questions. An important case, addressed here, is that of temporal questions, where cues for temporal relations need to be discovered and handled. We present TEQUILA, an enabler method for temporal QA that can run on top of any KB-QA engine. TEQUILA has four stages. It detects if a question has temporal intent. It decomposes and rewrites the question into non-temporal sub-questions and temporal constraints. Answers to sub-questions are then retrieved from the underlying KB-QA engine. Finally, TEQUILA uses constraint reasoning on temporal intervals to compute final answers to the full question. Comparisons against state-of-the-art baselines show the viability of our method.
Tasks Question Answering
Published 2019-08-09
URL https://arxiv.org/abs/1908.03650v3
PDF https://arxiv.org/pdf/1908.03650v3.pdf
PWC https://paperswithcode.com/paper/tequila-temporal-question-answering-over
Repo
Framework

Bidding in Spades

Title Bidding in Spades
Authors Gal Cohensius, Reshef Meir, Nadav Oved, Roni Stern
Abstract We present a Spades bidding algorithm that is superior to recreational human players and to publicly available bots. Like in Bridge, the game of Spades is composed of two independent phases, \textit{bidding} and \textit{playing}. This paper focuses on the bidding algorithm, since this phase holds a precise challenge: based on the input, choose the bid that maximizes the agent’s winning probability. Our \emph{Bidding-in-Spades} (BIS) algorithm heuristically determines the bidding strategy by comparing the expected utility of each possible bid. A major challenge is how to estimate these expected utilities. To this end, we propose a set of domain-specific heuristics, and then correct them via machine learning using data from real-world players. The \BIS algorithm we present can be attached to any playing algorithm. It beats rule-based bidding bots when all use the same playing component. When combined with a rule-based playing algorithm, it is superior to the average recreational human.
Tasks
Published 2019-12-24
URL https://arxiv.org/abs/1912.11323v2
PDF https://arxiv.org/pdf/1912.11323v2.pdf
PWC https://paperswithcode.com/paper/bidding-in-spades
Repo
Framework

Repeated A/B Testing

Title Repeated A/B Testing
Authors Nicolò Cesa-Bianchi, Tommaso R. Cesari, Yishay Mansour, Vianney Perchet
Abstract A/B testing is of central importance in the industry, especially in the technology sector where companies run long sequences of A/B tests to optimize their products. Since the space of potential innovations is typically vast, the experimenter must make quick and good decisions without wasting too much time on a single A/B test in the sequence. In particular, discarding an innovation with a small benefit might be better in the long run than using many samples to precisely determine its value. In this work, we introduce a performance measure that captures this idea and design an efficient algorithm that performs almost as well as the best A/B strategy in a given set. As it turns out, a key technical difficulty that significantly affects the learning rates is the hardness of obtaining unbiased estimates of the strategy rewards.
Tasks
Published 2019-05-28
URL https://arxiv.org/abs/1905.11797v3
PDF https://arxiv.org/pdf/1905.11797v3.pdf
PWC https://paperswithcode.com/paper/repeated-ab-testing
Repo
Framework

Using a Segmenting Description approach in Multiple Criteria Decision Aiding

Title Using a Segmenting Description approach in Multiple Criteria Decision Aiding
Authors Milosz Kadzinski, Jan Badura, Jose Rui Figueira
Abstract We propose a new method for analyzing a set of parameters in a multiple criteria ranking method. Unlike the existing techniques, we do not use any optimization technique, instead incorporating and extending a Segmenting Description approach. While considering a value-based preference disaggregation method, we demonstrate the usefulness of the introduced algorithm in a multi-purpose decision analysis exploiting a system of inequalities that models the Decision Maker’s preferences. Specifically, we discuss how it can be applied for verifying the consistency between the revealed and estimated preferences as well as for identifying the sources of potential incoherence. Moreover, we employ the method for conducting robustness analysis, i.e., discovering a set of all compatible parameter values and verifying the stability of suggested recommendation in view of multiplicity of feasible solutions. In addition, we make clear its suitability for generating arguments about the validity of outcomes and the role of particular criteria. We discuss the favorable characteristics of the Segmenting Description approach which enhance its suitability for use in Multiple Criteria Decision Aiding. These include keeping in memory an entire process of transforming a system of inequalities and avoiding the need for processing the inequalities contained in the basic system which is subsequently enriched with some hypothesis to be verified. The applicability of the proposed method is exemplified on a numerical study.
Tasks
Published 2019-03-05
URL http://arxiv.org/abs/1903.01923v1
PDF http://arxiv.org/pdf/1903.01923v1.pdf
PWC https://paperswithcode.com/paper/using-a-segmenting-description-approach-in
Repo
Framework

On the Power and Limitations of Random Features for Understanding Neural Networks

Title On the Power and Limitations of Random Features for Understanding Neural Networks
Authors Gilad Yehudai, Ohad Shamir
Abstract Recently, a spate of papers have provided positive theoretical results for training over-parameterized neural networks (where the network size is larger than what is needed to achieve low error). The key insight is that with sufficient over-parameterization, gradient-based methods will implicitly leave some components of the network relatively unchanged, so the optimization dynamics will behave as if those components are essentially fixed at their initial random values. In fact, fixing these explicitly leads to the well-known approach of learning with random features. In other words, these techniques imply that we can successfully learn with neural networks, whenever we can successfully learn with random features. In this paper, we first review these techniques, providing a simple and self-contained analysis for one-hidden-layer networks. We then argue that despite the impressive positive results, random feature approaches are also inherently limited in what they can explain. In particular, we rigorously show that random features cannot be used to learn even a single ReLU neuron with standard Gaussian inputs, unless the network size (or magnitude of the weights) is exponentially large. Since a single neuron is learnable with gradient-based methods, we conclude that we are still far from a satisfying general explanation for the empirical success of neural networks.
Tasks
Published 2019-04-01
URL https://arxiv.org/abs/1904.00687v3
PDF https://arxiv.org/pdf/1904.00687v3.pdf
PWC https://paperswithcode.com/paper/on-the-power-and-limitations-of-random
Repo
Framework

Single-Channel Speech Separation with Auxiliary Speaker Embeddings

Title Single-Channel Speech Separation with Auxiliary Speaker Embeddings
Authors Shuo Liu, Gil Keren, Björn Schuller
Abstract We present a novel source separation model to decompose asingle-channel speech signal into two speech segments belonging to two different speakers. The proposed model is a neural network based on residual blocks, and uses learnt speaker embeddings created from additional clean context recordings of the two speakers as input to assist in attributing the different time-frequency bins to the two speakers. In experiments, we show that the proposed model yields good performance in the source separation task, and outperforms the state-of-the-art baselines. Specifically, separating speech from the challenging VoxCeleb dataset, the proposed model yields 4.79dB signal-to-distortion ratio, 8.44dB signal-to-artifacts ratio and 7.11dB signal-to-interference ratio.
Tasks Speech Separation
Published 2019-06-24
URL https://arxiv.org/abs/1906.09997v1
PDF https://arxiv.org/pdf/1906.09997v1.pdf
PWC https://paperswithcode.com/paper/single-channel-speech-separation-with
Repo
Framework

Fast Hierarchical Neural Network for Feature Learning on Point Cloud

Title Fast Hierarchical Neural Network for Feature Learning on Point Cloud
Authors Can Chen, Luca Zanotti Fragonara, Antonios Tsourdos
Abstract The analyses relying on 3D point clouds are an utterly complex task, often involving million of points, but also requiring computationally efficient algorithms because of many real-time applications; e.g. autonomous vehicle. However, point clouds are intrinsically irregular and the points are sparsely distributed in a non-Euclidean space, which normally requires point-wise processing to achieve high performances. Although shared filter matrices and pooling layers in convolutional neural networks (CNNs) are capable of reducing the dimensionality of the problem and extracting high-level information simultaneously, grids and highly regular data format are required as input. In order to balance model performance and complexity, we introduce a novel neural network architecture exploiting local features from a manually subsampled point set. In our network, a recursive farthest point sampling method is firstly applied to efficiently cover the entire point set. Successively, we employ the k-nearest neighbours (knn) algorithm to gather local neighbourhood for each group of the subsampled points. Finally, a multiple layer perceptron (MLP) is applied on the subsampled points and edges that connect corresponding point and neighbours to extract local features. The architecture has been tested for both shape classification and segmentation using the ModelNet40 and ShapeNet part datasets, in order to show that the network achieves the best trade-off in terms of competitive performance when compared to other state-of-the-art algorithms.
Tasks
Published 2019-06-10
URL https://arxiv.org/abs/1906.04117v1
PDF https://arxiv.org/pdf/1906.04117v1.pdf
PWC https://paperswithcode.com/paper/fast-hierarchical-neural-network-for-feature
Repo
Framework

Fast and Scalable Estimator for Sparse and Unit-Rank Higher-Order Regression Models

Title Fast and Scalable Estimator for Sparse and Unit-Rank Higher-Order Regression Models
Authors Jiaqi Zhang, Beilun Wang
Abstract Because tensor data appear more and more frequently in various scientific researches and real-world applications, analyzing the relationship between tensor features and the univariate outcome becomes an elementary task in many fields. To solve this task, we propose \underline{Fa}st \underline{S}parse \underline{T}ensor \underline{R}egression model (FasTR) based on so-called unit-rank CANDECOMP/PARAFAC decomposition. FasTR first decomposes the tensor coefficient into component vectors and then estimates each vector with $\ell_1$ regularized regression. Because of the independence of component vectors, FasTR is able to solve in a parallel way and the time complexity is proved to be superior to previous models. We evaluate the performance of FasTR on several simulated datasets and a real-world fMRI dataset. Experiment results show that, compared with four baseline models, in every case, FasTR can compute a better solution within less time.
Tasks
Published 2019-11-29
URL https://arxiv.org/abs/1912.01450v1
PDF https://arxiv.org/pdf/1912.01450v1.pdf
PWC https://paperswithcode.com/paper/fast-and-scalable-estimator-for-sparse-and
Repo
Framework

The Role of Cooperation in Responsible AI Development

Title The Role of Cooperation in Responsible AI Development
Authors Amanda Askell, Miles Brundage, Gillian Hadfield
Abstract In this paper, we argue that competitive pressures could incentivize AI companies to underinvest in ensuring their systems are safe, secure, and have a positive social impact. Ensuring that AI systems are developed responsibly may therefore require preventing and solving collective action problems between companies. We note that there are several key factors that improve the prospects for cooperation in collective action problems. We use this to identify strategies to improve the prospects for industry cooperation on the responsible development of AI.
Tasks
Published 2019-07-10
URL https://arxiv.org/abs/1907.04534v1
PDF https://arxiv.org/pdf/1907.04534v1.pdf
PWC https://paperswithcode.com/paper/the-role-of-cooperation-in-responsible-ai
Repo
Framework
comments powered by Disqus