Paper Group AWR 64
Detecting Overfitting of Deep Generative Networks via Latent Recovery. Training Neural Response Selection for Task-Oriented Dialogue Systems. Interpreting and improving natural-language processing (in machines) with natural language-processing (in the brain). Scalable Gaussian Process Regression for Kernels with a Non-Stationary Phase. Planted Hitt …
Detecting Overfitting of Deep Generative Networks via Latent Recovery
Title | Detecting Overfitting of Deep Generative Networks via Latent Recovery |
Authors | Ryan Webster, Julien Rabin, Loic Simon, Frederic Jurie |
Abstract | State of the art deep generative networks are capable of producing images with such incredible realism that they can be suspected of memorizing training images. It is why it is not uncommon to include visualizations of training set nearest neighbors, to suggest generated images are not simply memorized. We demonstrate this is not sufficient and motivates the need to study memorization/overfitting of deep generators with more scrutiny. This paper addresses this question by i) showing how simple losses are highly effective at reconstructing images for deep generators ii) analyzing the statistics of reconstruction errors when reconstructing training and validation images, which is the standard way to analyze overfitting in machine learning. Using this methodology, this paper shows that overfitting is not detectable in the pure GAN models proposed in the literature, in contrast with those using hybrid adversarial losses, which are amongst the most widely applied generative methods. The paper also shows that standard GAN evaluation metrics fail to capture memorization for some deep generators. Finally, the paper also shows how off-the-shelf GAN generators can be successfully applied to face inpainting and face super-resolution using the proposed reconstruction method, without hybrid adversarial losses. |
Tasks | Facial Inpainting, Super-Resolution |
Published | 2019-01-09 |
URL | http://arxiv.org/abs/1901.03396v1 |
http://arxiv.org/pdf/1901.03396v1.pdf | |
PWC | https://paperswithcode.com/paper/detecting-overfitting-of-deep-generative |
Repo | https://github.com/ryanwebster90/gen-overfitting-latent-recovery |
Framework | pytorch |
Training Neural Response Selection for Task-Oriented Dialogue Systems
Title | Training Neural Response Selection for Task-Oriented Dialogue Systems |
Authors | Matthew Henderson, Ivan Vulić, Daniela Gerz, Iñigo Casanueva, Paweł Budzianowski, Sam Coope, Georgios Spithourakis, Tsung-Hsien Wen, Nikola Mrkšić, Pei-Hao Su |
Abstract | Despite their popularity in the chatbot literature, retrieval-based models have had modest impact on task-oriented dialogue systems, with the main obstacle to their application being the low-data regime of most task-oriented dialogue tasks. Inspired by the recent success of pretraining in language modelling, we propose an effective method for deploying response selection in task-oriented dialogue. To train response selection models for task-oriented dialogue tasks, we propose a novel method which: 1) pretrains the response selection model on large general-domain conversational corpora; and then 2) fine-tunes the pretrained model for the target dialogue domain, relying only on the small in-domain dataset to capture the nuances of the given dialogue domain. Our evaluation on six diverse application domains, ranging from e-commerce to banking, demonstrates the effectiveness of the proposed training method. |
Tasks | Chatbot, Language Modelling, Task-Oriented Dialogue Systems |
Published | 2019-06-04 |
URL | https://arxiv.org/abs/1906.01543v2 |
https://arxiv.org/pdf/1906.01543v2.pdf | |
PWC | https://paperswithcode.com/paper/training-neural-response-selection-for-task |
Repo | https://github.com/avidale/arxivarius |
Framework | none |
Interpreting and improving natural-language processing (in machines) with natural language-processing (in the brain)
Title | Interpreting and improving natural-language processing (in machines) with natural language-processing (in the brain) |
Authors | Mariya Toneva, Leila Wehbe |
Abstract | Neural networks models for NLP are typically implemented without the explicit encoding of language rules and yet they are able to break one performance record after another. This has generated a lot of research interest in interpreting the representations learned by these networks. We propose here a novel interpretation approach that relies on the only processing system we have that does understand language: the human brain. We use brain imaging recordings of subjects reading complex natural text to interpret word and sequence embeddings from 4 recent NLP models - ELMo, USE, BERT and Transformer-XL. We study how their representations differ across layer depth, context length, and attention type. Our results reveal differences in the context-related representations across these models. Further, in the transformer models, we find an interaction between layer depth and context length, and between layer depth and attention type. We finally hypothesize that altering BERT to better align with brain recordings would enable it to also better understand language. Probing the altered BERT using syntactic NLP tasks reveals that the model with increased brain-alignment outperforms the original model. Cognitive neuroscientists have already begun using NLP networks to study the brain, and this work closes the loop to allow the interaction between NLP and cognitive neuroscience to be a true cross-pollination. |
Tasks | |
Published | 2019-05-28 |
URL | https://arxiv.org/abs/1905.11833v4 |
https://arxiv.org/pdf/1905.11833v4.pdf | |
PWC | https://paperswithcode.com/paper/interpreting-and-improving-natural-language |
Repo | https://github.com/mtoneva/brain_language_nlp |
Framework | pytorch |
Scalable Gaussian Process Regression for Kernels with a Non-Stationary Phase
Title | Scalable Gaussian Process Regression for Kernels with a Non-Stationary Phase |
Authors | Jan Graßhoff, Alexandra Jankowski, Philipp Rostalski |
Abstract | The application of Gaussian processes (GPs) to large data sets is limited due to heavy memory and computational requirements. A variety of methods has been proposed to enable scalability, one of which is to exploit structure in the kernel matrix. Previous methods, however, cannot easily deal with non-stationary processes. This paper presents an efficient GP framework, that extends structured kernel interpolation methods to GPs with a non-stationary phase. We particularly treat mixtures of non-stationary processes, which are commonly used in the context of separation problems e.g. in biomedical signal processing. Our approach employs multiple sets of non-equidistant inducing points to account for the non-stationarity and retrieve Toeplitz and Kronecker structure in the kernel matrix allowing for efficient inference. Kernel learning is done by optimizing the marginal likelihood, which can be approximated efficiently using stochastic trace estimation methods. Our approach is demonstrated on numerical examples and large biomedical datasets. |
Tasks | Gaussian Processes |
Published | 2019-12-25 |
URL | https://arxiv.org/abs/1912.11713v1 |
https://arxiv.org/pdf/1912.11713v1.pdf | |
PWC | https://paperswithcode.com/paper/scalable-gaussian-process-regression-for |
Repo | https://github.com/ime-luebeck/non-stationary-phase-gp-mod |
Framework | none |
Planted Hitting Set Recovery in Hypergraphs
Title | Planted Hitting Set Recovery in Hypergraphs |
Authors | Ilya Amburg, Jon Kleinberg, Austin R. Benson |
Abstract | In various application areas, networked data is collected by measuring interactions involving some specific set of core nodes. This results in a network dataset containing the core nodes along with a potentially much larger set of fringe nodes that all have at least one interaction with a core node. In many settings, this type of data arises for structures that are richer than graphs, because they involve the interactions of larger sets; for example, the core nodes might be a set of individuals under surveillance, where we observe the attendees of meetings involving at least one of the core individuals. We model such scenarios using hypergraphs, and we study the problem of core recovery: if we observe the hypergraph but not the labels of core and fringe nodes, can we recover the “planted” set of core nodes in the hypergraph? We provide a theoretical framework for analyzing the recovery of such a set of core nodes and use our theory to develop a practical and scalable algorithm for core recovery. The crux of our analysis and algorithm is that the core nodes are a hitting set of the hypergraph, meaning that every hyperedge has at least one node in the set of core nodes. We demonstrate the efficacy of our algorithm on a number of real-world datasets, outperforming competitive baselines derived from network centrality and core-periphery measures. |
Tasks | |
Published | 2019-05-14 |
URL | https://arxiv.org/abs/1905.05839v1 |
https://arxiv.org/pdf/1905.05839v1.pdf | |
PWC | https://paperswithcode.com/paper/planted-hitting-set-recovery-in-hypergraphs |
Repo | https://github.com/ilyaamburg/Hypergraph-Planted-Hitting-Set-Recovery |
Framework | none |
Invertible Gaussian Reparameterization: Revisiting the Gumbel-Softmax
Title | Invertible Gaussian Reparameterization: Revisiting the Gumbel-Softmax |
Authors | Andres Potapczynski, Gabriel Loaiza-Ganem, John P. Cunningham |
Abstract | The Gumbel-Softmax is a continuous distribution over the simplex that is often used as a relaxation of discrete distributions. Because it can be readily interpreted and easily reparameterized, it enjoys widespread use. Unfortunately, we show that the cost of this aesthetic interpretability is material: the temperature hyperparameter must be set too high, KL estimates are noisy, and as a result, performance suffers. We circumvent the previous issues by proposing a much simpler and more flexible reparameterizable family of distributions that transforms Gaussian noise into a one-hot approximation through an invertible function. This invertible function is composed of a modified softmax and can incorporate diverse transformations that serve different specific purposes. For example, the stick-breaking procedure allows us to extend the reparameterization trick to distributions with countably infinite support, or normalizing flows let us increase the flexibility of the distribution. Our construction improves numerical stability and outperforms the Gumbel-Softmax in a variety of experiments while generating samples that are closer to their discrete counterparts and achieving lower-variance gradients. |
Tasks | |
Published | 2019-12-19 |
URL | https://arxiv.org/abs/1912.09588v2 |
https://arxiv.org/pdf/1912.09588v2.pdf | |
PWC | https://paperswithcode.com/paper/invertible-gaussian-reparameterization |
Repo | https://github.com/cunningham-lab/igr |
Framework | tf |
Digital phase-only holography using deep conditional generative models
Title | Digital phase-only holography using deep conditional generative models |
Authors | Jannes Gladrow |
Abstract | Holographic wave-shaping has found numerous applications across the physical sciences, especially since the development of digital spatial-light modulators (SLMs). A key challenge in digital holography consists in finding optimal hologram patterns which transform the incoming laser beam into desired shapes in a conjugate optical plane. The existing repertoire of approaches to solve this inverse problem is built on iterative phase-retrieval algorithms, which do not take optical aberrations and deviations from theoretical models into account. Here, we adopt a physics-free, data-driven, and probabilistic approach to the problem. Using deep conditional generative models such as Generative-Adversarial Networks (cGAN) or Variational Autoencoder (cVAE), we approximate conditional distributions of holograms for a given target laser intensity pattern. In order to reduce the cardinality of the problem, we train our models on a proxy mapping relating an 8x8-matrix of complex-valued spatial-frequency coefficients to the ensuing 100x100-shaped intensity distribution recorded on a camera. We discuss the degree of ‘ill-posedness’ that remains in this reduced problem and compare different generative model architectures in terms of their ability to find holograms that reconstruct given intensity patterns. Finally, we challenge our models to generalise to synthetic target intensities, where the existence of matching holograms cannot be guaranteed. We devise a forward-interpolating training scheme aimed at providing models the ability to interpolate in laser intensity space, rather than hologram space and show that this indeed enhances model performance on synthetic data sets. |
Tasks | |
Published | 2019-11-03 |
URL | https://arxiv.org/abs/1911.00904v1 |
https://arxiv.org/pdf/1911.00904v1.pdf | |
PWC | https://paperswithcode.com/paper/digital-phase-only-holography-using-deep |
Repo | https://github.com/JamesGlare/Holo_gen_models |
Framework | tf |
PolarMask: Single Shot Instance Segmentation with Polar Representation
Title | PolarMask: Single Shot Instance Segmentation with Polar Representation |
Authors | Enze Xie, Peize Sun, Xiaoge Song, Wenhai Wang, Ding Liang, Chunhua Shen, Ping Luo |
Abstract | In this paper, we introduce an anchor-box free and single shot instance segmentation method, which is conceptually simple, fully convolutional and can be used as a mask prediction module for instance segmentation, by easily embedding it into most off-the-shelf detection methods. Our method, termed PolarMask, formulates the instance segmentation problem as instance center classification and dense distance regression in a polar coordinate. Moreover, we propose two effective approaches to deal with sampling high-quality center examples and optimization for dense distance regression, respectively, which can significantly improve the performance and simplify the training process. Without any bells and whistles, PolarMask achieves 32.9% in mask mAP with single-model and single-scale training/testing on challenging COCO dataset. For the first time, we demonstrate a much simpler and flexible instance segmentation framework achieving competitive accuracy. We hope that the proposed PolarMask framework can serve as a fundamental and strong baseline for single shot instance segmentation tasks. Code is available at: github.com/xieenze/PolarMask. |
Tasks | Instance Segmentation, Object Detection, Semantic Segmentation |
Published | 2019-09-29 |
URL | https://arxiv.org/abs/1909.13226v4 |
https://arxiv.org/pdf/1909.13226v4.pdf | |
PWC | https://paperswithcode.com/paper/polarmask-single-shot-instance-segmentation |
Repo | https://github.com/xieenze/PolarMask |
Framework | pytorch |
EmbedMask: Embedding Coupling for One-stage Instance Segmentation
Title | EmbedMask: Embedding Coupling for One-stage Instance Segmentation |
Authors | Hui Ying, Zhaojin Huang, Shu Liu, Tianjia Shao, Kun Zhou |
Abstract | Current instance segmentation methods can be categorized into segmentation-based methods that segment first then do clustering, and proposal-based methods that detect first then predict masks for each instance proposal using repooling. In this work, we propose a one-stage method, named EmbedMask, that unifies both methods by taking advantages of them. Like proposal-based methods, EmbedMask builds on top of detection models making it strong in detection capability. Meanwhile, EmbedMask applies extra embedding modules to generate embeddings for pixels and proposals, where pixel embeddings are guided by proposal embeddings if they belong to the same instance. Through this embedding coupling process, pixels are assigned to the mask of the proposal if their embeddings are similar. The pixel-level clustering enables EmbedMask to generate high-resolution masks without missing details from repooling, and the existence of proposal embedding simplifies and strengthens the clustering procedure to achieve high speed with higher performance than segmentation-based methods. Without any bells and whistles, EmbedMask achieves comparable performance as Mask R-CNN, which is the representative two-stage method, and can produce more detailed masks at a higher speed. Code is available at github.com/yinghdb/EmbedMask. |
Tasks | Instance Segmentation, Semantic Segmentation |
Published | 2019-12-04 |
URL | https://arxiv.org/abs/1912.01954v2 |
https://arxiv.org/pdf/1912.01954v2.pdf | |
PWC | https://paperswithcode.com/paper/embedmask-embedding-coupling-for-one-stage |
Repo | https://github.com/yinghdb/EmbedMask |
Framework | pytorch |
Parallel fault-tolerant programming of an arbitrary feedforward photonic network
Title | Parallel fault-tolerant programming of an arbitrary feedforward photonic network |
Authors | Sunil Pai, Ian A. D. Williamson, Tyler W. Hughes, Momchil Minkov, Olav Solgaard, Shanhui Fan, David A. B. Miller |
Abstract | Reconfigurable photonic mesh networks of tunable beamsplitter nodes can linearly transform $N$-dimensional vectors representing input modal amplitudes of light for applications such as energy-efficient machine learning hardware, quantum information processing, and mode demultiplexing. Such photonic meshes are typically programmed and/or calibrated by tuning or characterizing each beam splitter one-by-one, which can be time-consuming and can limit scaling to larger meshes. Here we introduce a graph-topological approach that defines the general class of feedforward networks commonly used in such applications and identifies columns of non-interacting nodes that can be adjusted simultaneously. By virtue of this approach, we can calculate the necessary input vectors to program entire columns of nodes in parallel by simultaneously nullifying the power in one output of each node via optoelectronic feedback onto adjustable phase shifters or couplers. This parallel nullification approach is fault-tolerant to fabrication errors, requiring no prior knowledge or calibration of the node parameters, and can reduce the programming time by a factor of order $N$ to being proportional to the optical depth (or number of node columns in the device). As a demonstration, we simulate our programming protocol on a feedforward optical neural network model trained to classify handwritten digit images from the MNIST dataset with up to 98% validation accuracy. |
Tasks | Calibration |
Published | 2019-09-11 |
URL | https://arxiv.org/abs/1909.06179v1 |
https://arxiv.org/pdf/1909.06179v1.pdf | |
PWC | https://paperswithcode.com/paper/parallel-fault-tolerant-programming-of-an |
Repo | https://github.com/solgaardlab/neurophox |
Framework | tf |
PixelHop: A Successive Subspace Learning (SSL) Method for Object Classification
Title | PixelHop: A Successive Subspace Learning (SSL) Method for Object Classification |
Authors | Yueru Chen, C. -C. Jay Kuo |
Abstract | A new machine learning methodology, called successive subspace learning (SSL), is introduced in this work. SSL contains four key ingredients: 1) successive near-to-far neighborhood expansion; 2) unsupervised dimension reduction via subspace approximation; 3) supervised dimension reduction via label-assisted regression (LAG); and 4) feature concatenation and decision making. An image-based object classification method, called PixelHop, is proposed to illustrate the SSL design. It is shown by experimental results that the PixelHop method outperforms the classic CNN model of similar model complexity in three benchmarking datasets (MNIST, Fashion MNIST and CIFAR-10). Although SSL and deep learning (DL) have some high-level concept in common, they are fundamentally different in model formulation, the training process and training complexity. Extensive discussion on the comparison of SSL and DL is made to provide further insights into the potential of SSL. |
Tasks | Decision Making, Dimensionality Reduction, Object Classification |
Published | 2019-09-17 |
URL | https://arxiv.org/abs/1909.08190v1 |
https://arxiv.org/pdf/1909.08190v1.pdf | |
PWC | https://paperswithcode.com/paper/pixelhop-a-successive-subspace-learning-ssl |
Repo | https://github.com/yifan-fanyi/Pixelhop |
Framework | none |
“Does 4-4-2 exist?” – An Analytics Approach to Understand and Classify Football Team Formations in Single Match Situations
Title | “Does 4-4-2 exist?” – An Analytics Approach to Understand and Classify Football Team Formations in Single Match Situations |
Authors | Eric Müller-Budack, Jonas Theiner, Robert Rein, Ralph Ewerth |
Abstract | The chances to win a football match can be significantly increased if the right tactic is chosen and the behavior of the opposite team is well anticipated. For this reason, every professional football club employs a team of game analysts. However, at present game performance analysis is done manually and therefore highly time-consuming. Consequently, automated tools to support the analysis process are required. In this context, one of the main tasks is to summarize team formations by patterns such as 4-4-2. In this paper, we introduce an analytics approach that automatically classifies and visualizes the team formation based on the players’ position data. We focus on single match situations instead of complete halftimes or matches to provide a more detailed analysis. A detailed analysis of individual match situations depending on ball possession and match segment length is provided. For this purpose, a visual summary is utilized that summarizes the team formation in a match segment. An expert annotation study is conducted that demonstrates 1) the complexity of the task and 2) the usefulness of the visualization of single situations to understand team formations. The suggested classification approach outperforms existing methods for formation classification. In particular, our approach gives insights about the shortcomings of using patterns like 4-4-2 to describe team formations. |
Tasks | |
Published | 2019-09-02 |
URL | https://arxiv.org/abs/1910.00412v1 |
https://arxiv.org/pdf/1910.00412v1.pdf | |
PWC | https://paperswithcode.com/paper/does-4-4-2-exist-an-analytics-approach-to |
Repo | https://github.com/naokiwifruit/classify-football-formation |
Framework | none |
Causal inference using Bayesian non-parametric quasi-experimental design
Title | Causal inference using Bayesian non-parametric quasi-experimental design |
Authors | Max Hinne, Marcel A. J. van Gerven, Luca Ambrogioni |
Abstract | The de facto standard for causal inference is the randomized controlled trial, where one compares an manipulated group with a control group in order to determine the effect of an intervention. However, this research design is not always realistically possible due to pragmatic or ethical concerns. In these situations, quasi-experimental designs may provide a solution, as these allow for causal conclusions at the cost of additional design assumptions. In this paper, we provide a generic framework for quasi-experimental design using Bayesian model comparison, and we show how it can be used as an alternative to several common research designs. We provide a theoretical motivation for a Gaussian process based approach and demonstrate its convenient use in a number of simulations. Finally, we apply the framework to determine the effect of population-based thresholds for municipality funding in France, of the 2005 smoking ban in Sicily on the number of acute coronary events, and of the effect of an alleged historical phantom border in the Netherlands on Dutch voting behaviour. |
Tasks | Causal Inference |
Published | 2019-11-15 |
URL | https://arxiv.org/abs/1911.06722v1 |
https://arxiv.org/pdf/1911.06722v1.pdf | |
PWC | https://paperswithcode.com/paper/causal-inference-using-bayesian-non |
Repo | https://github.com/mhinne/BNQD |
Framework | none |
Model-based clustering in very high dimensions via adaptive projections
Title | Model-based clustering in very high dimensions via adaptive projections |
Authors | Bernd Taschler, Frank Dondelinger, Sach Mukherjee |
Abstract | Mixture models are a standard approach to dealing with heterogeneous data with non-i.i.d. structure. However, when the dimension $p$ is large relative to sample size $n$ and where either or both of means and covariances/graphical models may differ between the latent groups, mixture models face statistical and computational difficulties and currently available methods cannot realistically go beyond $p ! \sim ! 10^4$ or so. We propose an approach called Model-based Clustering via Adaptive Projections (MCAP). Instead of estimating mixtures in the original space, we work with a low-dimensional representation obtained by linear projection. The projection dimension itself plays an important role and governs a type of bias-variance tradeoff with respect to recovery of the relevant signals. MCAP sets the projection dimension automatically in a data-adaptive manner, using a proxy for the assignment risk. Combining a full covariance formulation with the adaptive projection allows detection of both mean and covariance signals in very high dimensional problems. We show real-data examples in which covariance signals are reliably detected in problems with $p ! \sim ! 10^4$ or more, and simulations going up to $p = 10^6$. In some examples, MCAP performs well even when the mean signal is entirely removed, leaving differential covariance structure in the high-dimensional space as the only signal. Across a number of regimes, MCAP performs as well or better than a range of existing methods, including a recently-proposed $\ell_1$-penalized approach; and performance remains broadly stable with increasing dimension. MCAP can be run “out of the box” and is fast enough for interactive use on large-$p$ problems using standard desktop computing resources. |
Tasks | |
Published | 2019-02-22 |
URL | http://arxiv.org/abs/1902.08472v1 |
http://arxiv.org/pdf/1902.08472v1.pdf | |
PWC | https://paperswithcode.com/paper/model-based-clustering-in-very-high |
Repo | https://github.com/btaschler/mcap |
Framework | none |
Inexact Block Coordinate Descent Algorithms for Nonsmooth Nonconvex Optimization
Title | Inexact Block Coordinate Descent Algorithms for Nonsmooth Nonconvex Optimization |
Authors | Yang Yang, Marius Pesavento, Zhi-Quan Luo, Björn Ottersten |
Abstract | In this paper, we propose an inexact block coordinate descent algorithm for large-scale nonsmooth nonconvex optimization problems. At each iteration, a particular block variable is selected and updated by inexactly solving the original optimization problem with respect to that block variable. More precisely, a local approximation of the original optimization problem is solved. The proposed algorithm has several attractive features, namely, i) high flexibility, as the approximation function only needs to be strictly convex and it does not have to be a global upper bound of the original function; ii) fast convergence, as the approximation function can be designed to exploit the problem structure at hand and the stepsize is calculated by the line search; iii) low complexity, as the approximation subproblems are much easier to solve and the line search scheme is carried out over a properly constructed differentiable function; iv) guaranteed convergence of a subsequence to a stationary point, even when the objective function does not have a Lipschitz continuous gradient. Interestingly, when the approximation subproblem is solved by a descent algorithm, convergence of a subsequence to a stationary point is still guaranteed even if the approximation subproblem is solved inexactly by terminating the descent algorithm after a finite number of iterations. These features make the proposed algorithm suitable for large-scale problems where the dimension exceeds the memory and/or the processing capability of the existing hardware. These features are also illustrated by several applications in signal processing and machine learning, for instance, network anomaly detection and phase retrieval. |
Tasks | Anomaly Detection |
Published | 2019-05-10 |
URL | https://arxiv.org/abs/1905.04211v5 |
https://arxiv.org/pdf/1905.04211v5.pdf | |
PWC | https://paperswithcode.com/paper/inexact-block-coordinate-descent-algorithms |
Repo | https://github.com/optyang/BSCA |
Framework | none |