Paper Group AWR 293
Machine Learning Classifiers Do Not Improve the Prediction of Academic Risk: Evidence from Australia. Deep Reinforcement Learning for General Video Game AI. Links: A High-Dimensional Online Clustering Method. Hybrid Noise Removal in Hyperspectral Imagery With a Spatial-Spectral Gradient Network. I-HAZE: a dehazing benchmark with real hazy and haze- …
Machine Learning Classifiers Do Not Improve the Prediction of Academic Risk: Evidence from Australia
Title | Machine Learning Classifiers Do Not Improve the Prediction of Academic Risk: Evidence from Australia |
Authors | Sarah Cornell-Farrow, Robert Garrard |
Abstract | Machine learning methods tend to outperform traditional statistical models at prediction. In the prediction of academic achievement, ML models have not shown substantial improvement over logistic regression. So far, these results have almost entirely focused on college achievement, due to the availability of administrative datasets, and have contained relatively small sample sizes by ML standards. In this article we apply popular machine learning models to a large dataset ($n=1.2$ million) containing primary and middle school performance on a standardized test given annually to Australian students. We show that machine learning models do not outperform logistic regression for detecting students who will perform in the `below standard’ band of achievement upon sitting their next test, even in a large-$n$ setting. | |
Tasks | |
Published | 2018-07-19 |
URL | https://arxiv.org/abs/1807.07215v4 |
https://arxiv.org/pdf/1807.07215v4.pdf | |
PWC | https://paperswithcode.com/paper/machine-learning-classifiers-do-not-improve |
Repo | https://github.com/RobGarrard/A-Machine-Learning-Approach-for-Detecting-Students-at-Risk-of-Low-Academic-Achievement |
Framework | tf |
Deep Reinforcement Learning for General Video Game AI
Title | Deep Reinforcement Learning for General Video Game AI |
Authors | Ruben Rodriguez Torrado, Philip Bontrager, Julian Togelius, Jialin Liu, Diego Perez-Liebana |
Abstract | The General Video Game AI (GVGAI) competition and its associated software framework provides a way of benchmarking AI algorithms on a large number of games written in a domain-specific description language. While the competition has seen plenty of interest, it has so far focused on online planning, providing a forward model that allows the use of algorithms such as Monte Carlo Tree Search. In this paper, we describe how we interface GVGAI to the OpenAI Gym environment, a widely used way of connecting agents to reinforcement learning problems. Using this interface, we characterize how widely used implementations of several deep reinforcement learning algorithms fare on a number of GVGAI games. We further analyze the results to provide a first indication of the relative difficulty of these games relative to each other, and relative to those in the Arcade Learning Environment under similar conditions. |
Tasks | Atari Games |
Published | 2018-06-06 |
URL | http://arxiv.org/abs/1806.02448v1 |
http://arxiv.org/pdf/1806.02448v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-reinforcement-learning-for-general-video |
Repo | https://github.com/rubenrtorrado/GVGAI_GYM |
Framework | none |
Links: A High-Dimensional Online Clustering Method
Title | Links: A High-Dimensional Online Clustering Method |
Authors | Philip Andrew Mansfield, Quan Wang, Carlton Downey, Li Wan, Ignacio Lopez Moreno |
Abstract | We present a novel algorithm, called Links, designed to perform online clustering on unit vectors in a high-dimensional Euclidean space. The algorithm is appropriate when it is necessary to cluster data efficiently as it streams in, and is to be contrasted with traditional batch clustering algorithms that have access to all data at once. For example, Links has been successfully applied to embedding vectors generated from face images or voice recordings for the purpose of recognizing people, thereby providing real-time identification during video or audio capture. |
Tasks | |
Published | 2018-01-30 |
URL | http://arxiv.org/abs/1801.10123v1 |
http://arxiv.org/pdf/1801.10123v1.pdf | |
PWC | https://paperswithcode.com/paper/links-a-high-dimensional-online-clustering |
Repo | https://github.com/QEDan/links_clustering |
Framework | none |
Hybrid Noise Removal in Hyperspectral Imagery With a Spatial-Spectral Gradient Network
Title | Hybrid Noise Removal in Hyperspectral Imagery With a Spatial-Spectral Gradient Network |
Authors | Qiang Zhang, Qiangqiang Yuan, Jie Li, Xinxin Liu, Huanfeng Shen, Liangpei Zhang |
Abstract | The existence of hybrid noise in hyperspectral images (HSIs) severely degrades the data quality, reduces the interpretation accuracy of HSIs, and restricts the subsequent HSIs applications. In this paper, the spatial-spectral gradient network (SSGN) is presented for mixed noise removal in HSIs. The proposed method employs a spatial-spectral gradient learning strategy, in consideration of the unique spatial structure directionality of sparse noise and spectral differences with additional complementary information for better extracting intrinsic and deep features of HSIs. Based on a fully cascaded multi-scale convolutional network, SSGN can simultaneously deal with the different types of noise in different HSIs or spectra by the use of the same model. The simulated and real-data experiments undertaken in this study confirmed that the proposed SSGN performs better at mixed noise removal than the other state-of-the-art HSI denoising algorithms, in evaluation indices, visual assessments, and time consumption. |
Tasks | Denoising |
Published | 2018-10-01 |
URL | http://arxiv.org/abs/1810.00495v3 |
http://arxiv.org/pdf/1810.00495v3.pdf | |
PWC | https://paperswithcode.com/paper/hybrid-noise-removal-in-hyperspectral-imagery |
Repo | https://github.com/WHUQZhang/SSGN |
Framework | none |
I-HAZE: a dehazing benchmark with real hazy and haze-free indoor images
Title | I-HAZE: a dehazing benchmark with real hazy and haze-free indoor images |
Authors | Codruta O. Ancuti, Cosmin Ancuti, Radu Timofte, Christophe De Vleeschouwer |
Abstract | Image dehazing has become an important computational imaging topic in the recent years. However, due to the lack of ground truth images, the comparison of dehazing methods is not straightforward, nor objective. To overcome this issue we introduce a new dataset -named I-HAZE- that contains 35 image pairs of hazy and corresponding haze-free (ground-truth) indoor images. Different from most of the existing dehazing databases, hazy images have been generated using real haze produced by a professional haze machine. For easy color calibration and improved assessment of dehazing algorithms, each scene include a MacBeth color checker. Moreover, since the images are captured in a controlled environment, both haze-free and hazy images are captured under the same illumination conditions. This represents an important advantage of the I-HAZE dataset that allows us to objectively compare the existing image dehazing techniques using traditional image quality metrics such as PSNR and SSIM. |
Tasks | Calibration, Image Dehazing |
Published | 2018-04-13 |
URL | http://arxiv.org/abs/1804.05091v1 |
http://arxiv.org/pdf/1804.05091v1.pdf | |
PWC | https://paperswithcode.com/paper/i-haze-a-dehazing-benchmark-with-real-hazy |
Repo | https://github.com/inyong37/Vision |
Framework | tf |
Complexity Reduction in the Negotiation of New Lexical Conventions
Title | Complexity Reduction in the Negotiation of New Lexical Conventions |
Authors | William Schueller, Vittorio Loreto, Pierre-Yves Oudeyer |
Abstract | In the process of collectively inventing new words for new concepts in a population, conflicts can quickly become numerous, in the form of synonymy and homonymy. Remembering all of them could cost too much memory, and remembering too few may slow down the overall process. Is there an efficient behavior that could help balance the two? The Naming Game is a multi-agent computational model for the emergence of language, focusing on the negotiation of new lexical conventions, where a common lexicon self-organizes but going through a phase of high complexity. Previous work has been done on the control of complexity growth in this particular model, by allowing agents to actively choose what they talk about. However, those strategies were relying on ad hoc heuristics highly dependent on fine-tuning of parameters. We define here a new principled measure and a new strategy, based on the beliefs of each agent on the global state of the population. The measure does not rely on heavy computation, and is cognitively plausible. The new strategy yields an efficient control of complexity growth, along with a faster agreement process. Also, we show that short-term memory is enough to build relevant beliefs about the global lexicon. |
Tasks | |
Published | 2018-05-15 |
URL | http://arxiv.org/abs/1805.05631v2 |
http://arxiv.org/pdf/1805.05631v2.pdf | |
PWC | https://paperswithcode.com/paper/complexity-reduction-in-the-negotiation-of |
Repo | https://github.com/wschuell/notebooks_cogsci2018 |
Framework | none |
Medical Concept Embedding with Time-Aware Attention
Title | Medical Concept Embedding with Time-Aware Attention |
Authors | Xiangrui Cai, Jinyang Gao, Kee Yuan Ngiam, Beng Chin Ooi, Ying Zhang, Xiaojie Yuan |
Abstract | Embeddings of medical concepts such as medication, procedure and diagnosis codes in Electronic Medical Records (EMRs) are central to healthcare analytics. Previous work on medical concept embedding takes medical concepts and EMRs as words and documents respectively. Nevertheless, such models miss out the temporal nature of EMR data. On the one hand, two consecutive medical concepts do not indicate they are temporally close, but the correlations between them can be revealed by the time gap. On the other hand, the temporal scopes of medical concepts often vary greatly (e.g., \textit{common cold} and \textit{diabetes}). In this paper, we propose to incorporate the temporal information to embed medical codes. Based on the Continuous Bag-of-Words model, we employ the attention mechanism to learn a “soft” time-aware context window for each medical concept. Experiments on public and proprietary datasets through clustering and nearest neighbour search tasks demonstrate the effectiveness of our model, showing that it outperforms five state-of-the-art baselines. |
Tasks | |
Published | 2018-06-06 |
URL | http://arxiv.org/abs/1806.02873v1 |
http://arxiv.org/pdf/1806.02873v1.pdf | |
PWC | https://paperswithcode.com/paper/medical-concept-embedding-with-time-aware |
Repo | https://github.com/XiangruiCAI/mce |
Framework | none |
Sylvester Normalizing Flows for Variational Inference
Title | Sylvester Normalizing Flows for Variational Inference |
Authors | Rianne van den Berg, Leonard Hasenclever, Jakub M. Tomczak, Max Welling |
Abstract | Variational inference relies on flexible approximate posterior distributions. Normalizing flows provide a general recipe to construct flexible variational posteriors. We introduce Sylvester normalizing flows, which can be seen as a generalization of planar flows. Sylvester normalizing flows remove the well-known single-unit bottleneck from planar flows, making a single transformation much more flexible. We compare the performance of Sylvester normalizing flows against planar flows and inverse autoregressive flows and demonstrate that they compare favorably on several datasets. |
Tasks | |
Published | 2018-03-15 |
URL | http://arxiv.org/abs/1803.05649v2 |
http://arxiv.org/pdf/1803.05649v2.pdf | |
PWC | https://paperswithcode.com/paper/sylvester-normalizing-flows-for-variational |
Repo | https://github.com/riannevdberg/sylvester-flows |
Framework | pytorch |
Relational recurrent neural networks
Title | Relational recurrent neural networks |
Authors | Adam Santoro, Ryan Faulkner, David Raposo, Jack Rae, Mike Chrzanowski, Theophane Weber, Daan Wierstra, Oriol Vinyals, Razvan Pascanu, Timothy Lillicrap |
Abstract | Memory-based neural networks model temporal data by leveraging an ability to remember information for long periods. It is unclear, however, whether they also have an ability to perform complex relational reasoning with the information they remember. Here, we first confirm our intuitions that standard memory architectures may struggle at tasks that heavily involve an understanding of the ways in which entities are connected – i.e., tasks involving relational reasoning. We then improve upon these deficits by using a new memory module – a \textit{Relational Memory Core} (RMC) – which employs multi-head dot product attention to allow memories to interact. Finally, we test the RMC on a suite of tasks that may profit from more capable relational reasoning across sequential information, and show large gains in RL domains (e.g. Mini PacMan), program evaluation, and language modeling, achieving state-of-the-art results on the WikiText-103, Project Gutenberg, and GigaWord datasets. |
Tasks | Language Modelling, Relational Reasoning |
Published | 2018-06-05 |
URL | http://arxiv.org/abs/1806.01822v2 |
http://arxiv.org/pdf/1806.01822v2.pdf | |
PWC | https://paperswithcode.com/paper/relational-recurrent-neural-networks |
Repo | https://github.com/L0SG/relational-rnn-pytorch |
Framework | pytorch |
Disconnected Manifold Learning for Generative Adversarial Networks
Title | Disconnected Manifold Learning for Generative Adversarial Networks |
Authors | Mahyar Khayatkhoei, Ahmed Elgammal, Maneesh Singh |
Abstract | Natural images may lie on a union of disjoint manifolds rather than one globally connected manifold, and this can cause several difficulties for the training of common Generative Adversarial Networks (GANs). In this work, we first show that single generator GANs are unable to correctly model a distribution supported on a disconnected manifold, and investigate how sample quality, mode dropping and local convergence are affected by this. Next, we show how using a collection of generators can address this problem, providing new insights into the success of such multi-generator GANs. Finally, we explain the serious issues caused by considering a fixed prior over the collection of generators and propose a novel approach for learning the prior and inferring the necessary number of generators without any supervision. Our proposed modifications can be applied on top of any other GAN model to enable learning of distributions supported on disconnected manifolds. We conduct several experiments to illustrate the aforementioned shortcoming of GANs, its consequences in practice, and the effectiveness of our proposed modifications in alleviating these issues. |
Tasks | |
Published | 2018-06-03 |
URL | http://arxiv.org/abs/1806.00880v3 |
http://arxiv.org/pdf/1806.00880v3.pdf | |
PWC | https://paperswithcode.com/paper/disconnected-manifold-learning-for-generative |
Repo | https://github.com/mahyarkoy/dmgan_release |
Framework | tf |
Adversarial Inference for Multi-Sentence Video Description
Title | Adversarial Inference for Multi-Sentence Video Description |
Authors | Jae Sung Park, Marcus Rohrbach, Trevor Darrell, Anna Rohrbach |
Abstract | While significant progress has been made in the image captioning task, video description is still in its infancy due to the complex nature of video data. Generating multi-sentence descriptions for long videos is even more challenging. Among the main issues are the fluency and coherence of the generated descriptions, and their relevance to the video. Recently, reinforcement and adversarial learning based methods have been explored to improve the image captioning models; however, both types of methods suffer from a number of issues, e.g. poor readability and high redundancy for RL and stability issues for GANs. In this work, we instead propose to apply adversarial techniques during inference, designing a discriminator which encourages better multi-sentence video description. In addition, we find that a multi-discriminator “hybrid” design, where each discriminator targets one aspect of a description, leads to the best results. Specifically, we decouple the discriminator to evaluate on three criteria: 1) visual relevance to the video, 2) language diversity and fluency, and 3) coherence across sentences. Our approach results in more accurate, diverse, and coherent multi-sentence video descriptions, as shown by automatic as well as human evaluation on the popular ActivityNet Captions dataset. |
Tasks | Image Captioning, Video Description |
Published | 2018-12-13 |
URL | http://arxiv.org/abs/1812.05634v2 |
http://arxiv.org/pdf/1812.05634v2.pdf | |
PWC | https://paperswithcode.com/paper/adversarial-inference-for-multi-sentence |
Repo | https://github.com/jamespark3922/adv-inf |
Framework | pytorch |
Deep Triplet Ranking Networks for One-Shot Recognition
Title | Deep Triplet Ranking Networks for One-Shot Recognition |
Authors | Meng Ye, Yuhong Guo |
Abstract | Despite the breakthroughs achieved by deep learning models in conventional supervised learning scenarios, their dependence on sufficient labeled training data in each class prevents effective applications of these deep models in situations where labeled training instances for a subset of novel classes are very sparse – in the extreme case only one instance is available for each class. To tackle this natural and important challenge, one-shot learning, which aims to exploit a set of well labeled base classes to build classifiers for the new target classes that have only one observed instance per class, has recently received increasing attention from the research community. In this paper we propose a novel end-to-end deep triplet ranking network to perform one-shot learning. The proposed approach learns class universal image embeddings on the well labeled base classes under a triplet ranking loss, such that the instances from new classes can be categorized based on their similarity with the one-shot instances in the learned embedding space. Moreover, our approach can naturally incorporate the available one-shot instances from the new classes into the embedding learning process to improve the triplet ranking model. We conduct experiments on two popular datasets for one-shot learning. The results show the proposed approach achieves better performance than the state-of-the- art comparison methods. |
Tasks | One-Shot Learning |
Published | 2018-04-19 |
URL | http://arxiv.org/abs/1804.07275v1 |
http://arxiv.org/pdf/1804.07275v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-triplet-ranking-networks-for-one-shot |
Repo | https://github.com/Mushfequr-Rahman/Omniglot_baselines |
Framework | pytorch |
AlphaGAN: Generative adversarial networks for natural image matting
Title | AlphaGAN: Generative adversarial networks for natural image matting |
Authors | Sebastian Lutz, Konstantinos Amplianitis, Aljosa Smolic |
Abstract | We present the first generative adversarial network (GAN) for natural image matting. Our novel generator network is trained to predict visually appealing alphas with the addition of the adversarial loss from the discriminator that is trained to classify well-composited images. Further, we improve existing encoder-decoder architectures to better deal with the spatial localization issues inherited in convolutional neural networks (CNN) by using dilated convolutions to capture global context information without downscaling feature maps and losing spatial information. We present state-of-the-art results on the alphamatting online benchmark for the gradient error and give comparable results in others. Our method is particularly well suited for fine structures like hair, which is of great importance in practical matting applications, e.g. in film/TV production. |
Tasks | Image Matting |
Published | 2018-07-26 |
URL | http://arxiv.org/abs/1807.10088v1 |
http://arxiv.org/pdf/1807.10088v1.pdf | |
PWC | https://paperswithcode.com/paper/alphagan-generative-adversarial-networks-for |
Repo | https://github.com/CDOTAD/AlphaGAN-Matting |
Framework | pytorch |
DLOPT: Deep Learning Optimization Library
Title | DLOPT: Deep Learning Optimization Library |
Authors | Andrés Camero, Jamal Toutouh, Enrique Alba |
Abstract | Deep learning hyper-parameter optimization is a tough task. Finding an appropriate network configuration is a key to success, however most of the times this labor is roughly done. In this work we introduce a novel library to tackle this problem, the Deep Learning Optimization Library: DLOPT. We briefly describe its architecture and present a set of use examples. This is an open source project developed under the GNU GPL v3 license and it is freely available at https://github.com/acamero/dlopt |
Tasks | |
Published | 2018-07-10 |
URL | http://arxiv.org/abs/1807.03523v1 |
http://arxiv.org/pdf/1807.03523v1.pdf | |
PWC | https://paperswithcode.com/paper/dlopt-deep-learning-optimization-library |
Repo | https://github.com/acamero/dlopt |
Framework | none |
Lipschitz-Margin Training: Scalable Certification of Perturbation Invariance for Deep Neural Networks
Title | Lipschitz-Margin Training: Scalable Certification of Perturbation Invariance for Deep Neural Networks |
Authors | Yusuke Tsuzuku, Issei Sato, Masashi Sugiyama |
Abstract | High sensitivity of neural networks against malicious perturbations on inputs causes security concerns. To take a steady step towards robust classifiers, we aim to create neural network models provably defended from perturbations. Prior certification work requires strong assumptions on network structures and massive computational costs, and thus the range of their applications was limited. From the relationship between the Lipschitz constants and prediction margins, we present a computationally efficient calculation technique to lower-bound the size of adversarial perturbations that can deceive networks, and that is widely applicable to various complicated networks. Moreover, we propose an efficient training procedure that robustifies networks and significantly improves the provably guarded areas around data points. In experimental evaluations, our method showed its ability to provide a non-trivial guarantee and enhance robustness for even large networks. |
Tasks | |
Published | 2018-02-12 |
URL | http://arxiv.org/abs/1802.04034v3 |
http://arxiv.org/pdf/1802.04034v3.pdf | |
PWC | https://paperswithcode.com/paper/lipschitz-margin-training-scalable |
Repo | https://github.com/ytsmiling/lmt |
Framework | none |