Paper Group AWR 304
Federated Learning with Additional Mechanisms on Clients to Reduce Communication Costs. Only sparsity based loss function for learning representations. Towards Learning of Filter-Level Heterogeneous Compression of Convolutional Neural Networks. Learning Parallax Attention for Stereo Image Super-Resolution. adVAE: A self-adversarial variational auto …
Federated Learning with Additional Mechanisms on Clients to Reduce Communication Costs
Title | Federated Learning with Additional Mechanisms on Clients to Reduce Communication Costs |
Authors | Xin Yao, Tianchi Huang, Chenglei Wu, Rui-Xiao Zhang, Lifeng Sun |
Abstract | Federated learning (FL) enables on-device training over distributed networks consisting of a massive amount of modern smart devices, such as smartphones and IoT (Internet of Things) devices. However, the leading optimization algorithm in such settings, i.e., federated averaging (FedAvg), suffers from heavy communication costs and the inevitable performance drop, especially when the local data is distributed in a non-IID way. To alleviate this problem, we propose two potential solutions by introducing additional mechanisms to the on-device training. The first (FedMMD) is adopting a two-stream model with the MMD (Maximum Mean Discrepancy) constraint instead of a single model in vanilla FedAvg to be trained on devices. Experiments show that the proposed method outperforms baselines, especially in non-IID FL settings, with a reduction of more than 20% in required communication rounds. The second is FL with feature fusion (FedFusion). By aggregating the features from both the local and global models, we achieve higher accuracy at fewer communication costs. Furthermore, the feature fusion modules offer better initialization for newly incoming clients and thus speed up the process of convergence. Experiments in popular FL scenarios show that our FedFusion outperforms baselines in both accuracy and generalization ability while reducing the number of required communication rounds by more than 60%. |
Tasks | |
Published | 2019-08-16 |
URL | https://arxiv.org/abs/1908.05891v2 |
https://arxiv.org/pdf/1908.05891v2.pdf | |
PWC | https://paperswithcode.com/paper/federated-learning-with-additional-mechanisms |
Repo | https://github.com/thu-media/FedFusion |
Framework | pytorch |
Only sparsity based loss function for learning representations
Title | Only sparsity based loss function for learning representations |
Authors | Vivek Bakaraju, Kishore Reddy Konda |
Abstract | We study the emergence of sparse representations in neural networks. We show that in unsupervised models with regularization, the emergence of sparsity is the result of the input data samples being distributed along highly non-linear or discontinuous manifold. We also derive a similar argument for discriminatively trained networks and present experiments to support this hypothesis. Based on our study of sparsity, we introduce a new loss function which can be used as regularization term for models like autoencoders and MLPs. Further, the same loss function can also be used as a cost function for an unsupervised single-layered neural network model for learning efficient representations. |
Tasks | |
Published | 2019-03-07 |
URL | http://arxiv.org/abs/1903.02893v1 |
http://arxiv.org/pdf/1903.02893v1.pdf | |
PWC | https://paperswithcode.com/paper/only-sparsity-based-loss-function-for |
Repo | https://github.com/Vivek-B/OVR |
Framework | tf |
Towards Learning of Filter-Level Heterogeneous Compression of Convolutional Neural Networks
Title | Towards Learning of Filter-Level Heterogeneous Compression of Convolutional Neural Networks |
Authors | Yochai Zur, Chaim Baskin, Evgenii Zheltonozhskii, Brian Chmiel, Itay Evron, Alex M. Bronstein, Avi Mendelson |
Abstract | Recently, deep learning has become a de facto standard in machine learning with convolutional neural networks (CNNs) demonstrating spectacular success on a wide variety of tasks. However, CNNs are typically very demanding computationally at inference time. One of the ways to alleviate this burden on certain hardware platforms is quantization relying on the use of low-precision arithmetic representation for the weights and the activations. Another popular method is the pruning of the number of filters in each layer. While mainstream deep learning methods train the neural networks weights while keeping the network architecture fixed, the emerging neural architecture search (NAS) techniques make the latter also amenable to training. In this paper, we formulate optimal arithmetic bit length allocation and neural network pruning as a NAS problem, searching for the configurations satisfying a computational complexity budget while maximizing the accuracy. We use a differentiable search method based on the continuous relaxation of the search space proposed by Liu et al. (arXiv:1806.09055). We show, by grid search, that heterogeneous quantized networks suffer from a high variance which renders the benefit of the search questionable. For pruning, improvement over homogeneous cases is possible, but it is still challenging to find those configurations with the proposed method. The code is publicly available at https://github.com/yochaiz/Slimmable and https://github.com/yochaiz/darts-UNIQ |
Tasks | Network Pruning, Neural Architecture Search, Quantization |
Published | 2019-04-22 |
URL | https://arxiv.org/abs/1904.09872v4 |
https://arxiv.org/pdf/1904.09872v4.pdf | |
PWC | https://paperswithcode.com/paper/towards-learning-of-filter-level |
Repo | https://github.com/yochaiz/darts-UNIQ |
Framework | pytorch |
Learning Parallax Attention for Stereo Image Super-Resolution
Title | Learning Parallax Attention for Stereo Image Super-Resolution |
Authors | Longguang Wang, Yingqian Wang, Zhengfa Liang, Zaiping Lin, Jungang Yang, Wei An, Yulan Guo |
Abstract | Stereo image pairs can be used to improve the performance of super-resolution (SR) since additional information is provided from a second viewpoint. However, it is challenging to incorporate this information for SR since disparities between stereo images vary significantly. In this paper, we propose a parallax-attention stereo superresolution network (PASSRnet) to integrate the information from a stereo image pair for SR. Specifically, we introduce a parallax-attention mechanism with a global receptive field along the epipolar line to handle different stereo images with large disparity variations. We also propose a new and the largest dataset for stereo image SR (namely, Flickr1024). Extensive experiments demonstrate that the parallax-attention mechanism can capture correspondence between stereo images to improve SR performance with a small computational and memory cost. Comparative results show that our PASSRnet achieves the state-of-the-art performance on the Middlebury, KITTI 2012 and KITTI 2015 datasets. |
Tasks | Image Super-Resolution, Stereo Image Super-Resolution, Super-Resolution |
Published | 2019-03-14 |
URL | http://arxiv.org/abs/1903.05784v3 |
http://arxiv.org/pdf/1903.05784v3.pdf | |
PWC | https://paperswithcode.com/paper/learning-parallax-attention-for-stereo-image |
Repo | https://github.com/YingqianWang/Flickr1024 |
Framework | none |
adVAE: A self-adversarial variational autoencoder with Gaussian anomaly prior knowledge for anomaly detection
Title | adVAE: A self-adversarial variational autoencoder with Gaussian anomaly prior knowledge for anomaly detection |
Authors | Xuhong Wang, Ying Du, Shijie Lin, Ping Cui, Yuntian Shen, Yupu Yang |
Abstract | Recently, deep generative models have become increasingly popular in unsupervised anomaly detection. However, deep generative models aim at recovering the data distribution rather than detecting anomalies. Besides, deep generative models have the risk of overfitting training samples, which has disastrous effects on anomaly detection performance. To solve the above two problems, we propose a Self-adversarial Variational Autoencoder with a Gaussian anomaly prior assumption. We assume that both the anomalous and the normal prior distribution are Gaussian and have overlaps in the latent space. Therefore, a Gaussian transformer net T is trained to synthesize anomalous but near-normal latent variables. Keeping the original training objective of Variational Autoencoder, besides, the generator G tries to distinguish between the normal latent variables and the anomalous ones synthesized by T, and the encoder E is trained to discriminate whether the output of G is real. These new objectives we added not only give both G and E the ability to discriminate but also introduce additional regularization to prevent overfitting. Compared with the SOTA baselines, the proposed model achieves significant improvements in extensive experiments. Datasets and our model are available at a Github repository. |
Tasks | Anomaly Detection, Unsupervised Anomaly Detection |
Published | 2019-03-03 |
URL | https://arxiv.org/abs/1903.00904v3 |
https://arxiv.org/pdf/1903.00904v3.pdf | |
PWC | https://paperswithcode.com/paper/self-adversarial-variational-autoencoder-with |
Repo | https://github.com/YeongHyeon/adVAE |
Framework | tf |
DualVD: An Adaptive Dual Encoding Model for Deep Visual Understanding in Visual Dialogue
Title | DualVD: An Adaptive Dual Encoding Model for Deep Visual Understanding in Visual Dialogue |
Authors | Xiaoze Jiang, Jing Yu, Zengchang Qin, Yingying Zhuang, Xingxing Zhang, Yue Hu, Qi Wu |
Abstract | Different from Visual Question Answering task that requires to answer only one question about an image, Visual Dialogue involves multiple questions which cover a broad range of visual content that could be related to any objects, relationships or semantics. The key challenge in Visual Dialogue task is thus to learn a more comprehensive and semantic-rich image representation which may have adaptive attentions on the image for variant questions. In this research, we propose a novel model to depict an image from both visual and semantic perspectives. Specifically, the visual view helps capture the appearance-level information, including objects and their relationships, while the semantic view enables the agent to understand high-level visual semantics from the whole image to the local regions. Futhermore, on top of such multi-view image features, we propose a feature selection framework which is able to adaptively capture question-relevant information hierarchically in fine-grained level. The proposed method achieved state-of-the-art results on benchmark Visual Dialogue datasets. More importantly, we can tell which modality (visual or semantic) has more contribution in answering the current question by visualizing the gate values. It gives us insights in understanding of human cognition in Visual Dialogue. |
Tasks | Feature Selection, Question Answering, Visual Dialog, Visual Question Answering |
Published | 2019-11-17 |
URL | https://arxiv.org/abs/1911.07251v1 |
https://arxiv.org/pdf/1911.07251v1.pdf | |
PWC | https://paperswithcode.com/paper/dualvd-an-adaptive-dual-encoding-model-for |
Repo | https://github.com/JXZe/DualVD |
Framework | pytorch |
A parallel Fortran framework for neural networks and deep learning
Title | A parallel Fortran framework for neural networks and deep learning |
Authors | Milan Curcic |
Abstract | This paper describes neural-fortran, a parallel Fortran framework for neural networks and deep learning. It features a simple interface to construct feed-forward neural networks of arbitrary structure and size, several activation functions, and stochastic gradient descent as the default optimization algorithm. Neural-fortran also leverages the Fortran 2018 standard collective subroutines to achieve data-based parallelism on shared- or distributed-memory machines. First, I describe the implementation of neural networks with Fortran derived types, whole-array arithmetic, and collective sum and broadcast operations to achieve parallelism. Second, I demonstrate the use of neural-fortran in an example of recognizing hand-written digits from images. Finally, I evaluate the computational performance in both serial and parallel modes. Ease of use and computational performance are similar to an existing popular machine learning framework, making neural-fortran a viable candidate for further development and use in production. |
Tasks | |
Published | 2019-02-18 |
URL | http://arxiv.org/abs/1902.06714v2 |
http://arxiv.org/pdf/1902.06714v2.pdf | |
PWC | https://paperswithcode.com/paper/a-parallel-fortran-framework-for-neural |
Repo | https://github.com/milancurcic/neural-fortran-paper |
Framework | tf |
BINet: Multi-perspective Business Process Anomaly Classification
Title | BINet: Multi-perspective Business Process Anomaly Classification |
Authors | Timo Nolle, Stefan Luettgen, Alexander Seeliger, Max Mühlhäuser |
Abstract | In this paper, we introduce BINet, a neural network architecture for real-time multi-perspective anomaly detection in business process event logs. BINet is designed to handle both the control flow and the data perspective of a business process. Additionally, we propose a set of heuristics for setting the threshold of an anomaly detection algorithm automatically. We demonstrate that BINet can be used to detect anomalies in event logs not only on a case level but also on event attribute level. Finally, we demonstrate that a simple set of rules can be used to utilize the output of BINet for anomaly classification. We compare BINet to eight other state-of-the-art anomaly detection algorithms and evaluate their performance on an elaborate data corpus of 29 synthetic and 15 real-life event logs. BINet outperforms all other methods both on the synthetic as well as on the real-life datasets. |
Tasks | Anomaly Detection |
Published | 2019-02-08 |
URL | http://arxiv.org/abs/1902.03155v1 |
http://arxiv.org/pdf/1902.03155v1.pdf | |
PWC | https://paperswithcode.com/paper/binet-multi-perspective-business-process |
Repo | https://github.com/tnolle/binet |
Framework | tf |
Way Off-Policy Batch Deep Reinforcement Learning of Implicit Human Preferences in Dialog
Title | Way Off-Policy Batch Deep Reinforcement Learning of Implicit Human Preferences in Dialog |
Authors | Natasha Jaques, Asma Ghandeharioun, Judy Hanwen Shen, Craig Ferguson, Agata Lapedriza, Noah Jones, Shixiang Gu, Rosalind Picard |
Abstract | Most deep reinforcement learning (RL) systems are not able to learn effectively from off-policy data, especially if they cannot explore online in the environment. These are critical shortcomings for applying RL to real-world problems where collecting data is expensive, and models must be tested offline before being deployed to interact with the environment – e.g. systems that learn from human interaction. Thus, we develop a novel class of off-policy batch RL algorithms, which are able to effectively learn offline, without exploring, from a fixed batch of human interaction data. We leverage models pre-trained on data as a strong prior, and use KL-control to penalize divergence from this prior during RL training. We also use dropout-based uncertainty estimates to lower bound the target Q-values as a more efficient alternative to Double Q-Learning. The algorithms are tested on the problem of open-domain dialog generation – a challenging reinforcement learning problem with a 20,000-dimensional action space. Using our Way Off-Policy algorithm, we can extract multiple different reward functions post-hoc from collected human interaction data, and learn effectively from all of these. We test the real-world generalization of these systems by deploying them live to converse with humans in an open-domain setting, and demonstrate that our algorithm achieves significant improvements over prior methods in off-policy batch RL. |
Tasks | Q-Learning |
Published | 2019-06-30 |
URL | https://arxiv.org/abs/1907.00456v2 |
https://arxiv.org/pdf/1907.00456v2.pdf | |
PWC | https://paperswithcode.com/paper/way-off-policy-batch-deep-reinforcement |
Repo | https://github.com/natashamjaques/neural_chat |
Framework | pytorch |
Towards Empathic Deep Q-Learning
Title | Towards Empathic Deep Q-Learning |
Authors | Bart Bussmann, Jacqueline Heinerman, Joel Lehman |
Abstract | As reinforcement learning (RL) scales to solve increasingly complex tasks, interest continues to grow in the fields of AI safety and machine ethics. As a contribution to these fields, this paper introduces an extension to Deep Q-Networks (DQNs), called Empathic DQN, that is loosely inspired both by empathy and the golden rule (“Do unto others as you would have them do unto you”). Empathic DQN aims to help mitigate negative side effects to other agents resulting from myopic goal-directed behavior. We assume a setting where a learning agent coexists with other independent agents (who receive unknown rewards), where some types of reward (e.g. negative rewards from physical harm) may generalize across agents. Empathic DQN combines the typical (self-centered) value with the estimated value of other agents, by imagining (by its own standards) the value of it being in the other’s situation (by considering constructed states where both agents are swapped). Proof-of-concept results in two gridworld environments highlight the approach’s potential to decrease collateral harms. While extending Empathic DQN to complex environments is non-trivial, we believe that this first step highlights the potential of bridge-work between machine ethics and RL to contribute useful priors for norm-abiding RL agents. |
Tasks | Q-Learning |
Published | 2019-06-26 |
URL | https://arxiv.org/abs/1906.10918v1 |
https://arxiv.org/pdf/1906.10918v1.pdf | |
PWC | https://paperswithcode.com/paper/towards-empathic-deep-q-learning |
Repo | https://github.com/bartbussmann/EmpathicDQN |
Framework | none |
A Story of Two Streams: Reinforcement Learning Models from Human Behavior and Neuropsychiatry
Title | A Story of Two Streams: Reinforcement Learning Models from Human Behavior and Neuropsychiatry |
Authors | Baihan Lin, Guillermo Cecchi, Djallel Bouneffouf, Jenna Reinen, Irina Rish |
Abstract | Drawing an inspiration from behavioral studies of human decision making, we propose here a more general and flexible parametric framework for reinforcement learning that extends standard Q-learning to a two-stream model for processing positive and negative rewards, and allows to incorporate a wide range of reward-processing biases – an important component of human decision making which can help us better understand a wide spectrum of multi-agent interactions in complex real-world socioeconomic systems, as well as various neuropsychiatric conditions associated with disruptions in normal reward processing. From the computational perspective, we observe that the proposed Split-QL model and its clinically inspired variants consistently outperform standard Q-Learning and SARSA methods, as well as recently proposed Double Q-Learning approaches, on simulated tasks with particular reward distributions, a real-world dataset capturing human decision-making in gambling tasks, and the Pac-Man game in a lifelong learning setting across different reward stationarities. |
Tasks | Decision Making, Q-Learning, Recommendation Systems |
Published | 2019-06-21 |
URL | https://arxiv.org/abs/1906.11286v6 |
https://arxiv.org/pdf/1906.11286v6.pdf | |
PWC | https://paperswithcode.com/paper/reinforcement-learning-models-of-human |
Repo | https://github.com/doerlbh/mentalRL |
Framework | pytorch |
Efficient Model-free Reinforcement Learning in Metric Spaces
Title | Efficient Model-free Reinforcement Learning in Metric Spaces |
Authors | Zhao Song, Wen Sun |
Abstract | Model-free Reinforcement Learning (RL) algorithms such as Q-learning [Watkins, Dayan 92] have been widely used in practice and can achieve human level performance in applications such as video games [Mnih et al. 15]. Recently, equipped with the idea of optimism in the face of uncertainty, Q-learning algorithms [Jin, Allen-Zhu, Bubeck, Jordan 18] can be proven to be sample efficient for discrete tabular Markov Decision Processes (MDPs) which have finite number of states and actions. In this work, we present an efficient model-free Q-learning based algorithm in MDPs with a natural metric on the state-action space–hence extending efficient model-free Q-learning algorithms to continuous state-action space. Compared to previous model-based RL algorithms for metric spaces [Kakade, Kearns, Langford 03], our algorithm does not require access to a black-box planning oracle. |
Tasks | Q-Learning |
Published | 2019-05-01 |
URL | http://arxiv.org/abs/1905.00475v1 |
http://arxiv.org/pdf/1905.00475v1.pdf | |
PWC | https://paperswithcode.com/paper/efficient-model-free-reinforcement-learning-1 |
Repo | https://github.com/seanrsinclair/AdaptiveQLearning |
Framework | none |
Automatic vocal tract landmark localization from midsagittal MRI data
Title | Automatic vocal tract landmark localization from midsagittal MRI data |
Authors | Mohammad Eslami, Christiane Neuschaefer-Rube, Antoine Serrurier |
Abstract | The various speech sounds of a language are obtained by varying the shape and position of the articulators surrounding the vocal tract. Analyzing their variations is crucial for understanding speech production, diagnosing speech disorders and planning therapy. Identifying key anatomical landmarks of these structures on medical images is a pre-requisite for any quantitative analysis and the rising amount of data generated in the field calls for an automatic solution. The challenge lies in the high inter- and intra-speaker variability, the mutual interaction between the articulators and the moderate quality of the images. This study addresses this issue for the first time and tackles it by means by means of Deep Learning. It proposes a dedicated network architecture named Flat-net and its performance are evaluated and compared with eleven state-of-the-art methods from the literature. The dataset contains midsagittal anatomical Magnetic Resonance Images for 9 speakers sustaining 62 articulations with 21 annotated anatomical landmarks per image. Results show that the Flat-net approach outperforms the former methods, leading to an overall Root Mean Square Error of 3.6 pixels/0.36 cm obtained in a leave-one-out procedure over the speakers. The implementation codes are also shared publicly on GitHub. |
Tasks | Face Alignment, Pose Estimation |
Published | 2019-07-18 |
URL | https://arxiv.org/abs/1907.07951v2 |
https://arxiv.org/pdf/1907.07951v2.pdf | |
PWC | https://paperswithcode.com/paper/automatic-vocal-tract-landmark-localization |
Repo | https://github.com/mohaEs/Train-Predict-Landmarks-by-flat-net |
Framework | tf |
Transfer Learning for Image-Based Malware Classification
Title | Transfer Learning for Image-Based Malware Classification |
Authors | Niket Bhodia, Pratikkumar Prajapati, Fabio Di Troia, Mark Stamp |
Abstract | In this paper, we consider the problem of malware detection and classification based on image analysis. We convert executable files to images and apply image recognition using deep learning (DL) models. To train these models, we employ transfer learning based on existing DL models that have been pre-trained on massive image datasets. We carry out various experiments with this technique and compare its performance to that of an extremely simple machine learning technique, namely, k-nearest neighbors (\kNN). For our k-NN experiments, we use features extracted directly from executables, rather than image analysis. While our image-based DL technique performs well in the experiments, surprisingly, it is outperformed by k-NN. We show that DL models are better able to generalize the data, in the sense that they outperform k-NN in simulated zero-day experiments. |
Tasks | Malware Classification, Malware Detection, Transfer Learning |
Published | 2019-01-21 |
URL | http://arxiv.org/abs/1903.11551v1 |
http://arxiv.org/pdf/1903.11551v1.pdf | |
PWC | https://paperswithcode.com/paper/transfer-learning-for-image-based-malware |
Repo | https://github.com/pratikpv/malware_classification |
Framework | none |
Estimating Attention Flow in Online Video Networks
Title | Estimating Attention Flow in Online Video Networks |
Authors | Siqi Wu, Marian-Andrei Rizoiu, Lexing Xie |
Abstract | Online videos have shown tremendous increase in Internet traffic. Most video hosting sites implement recommender systems, which connect the videos into a directed network and conceptually act as a source of pathways for users to navigate. At present, little is known about how human attention is allocated over such large-scale networks, and about the impacts of the recommender systems. In this paper, we first construct the Vevo network – a YouTube video network with 60,740 music videos interconnected by the recommendation links, and we collect their associated viewing dynamics. This results in a total of 310 million views every day over a period of 9 weeks. Next, we present large-scale measurements that connect the structure of the recommendation network and the video attention dynamics. We use the bow-tie structure to characterize the Vevo network and we find that its core component (23.1% of the videos), which occupies most of the attention (82.6% of the views), is made out of videos that are mainly recommended among themselves. This is indicative of the links between video recommendation and the inequality of attention allocation. Finally, we address the task of estimating the attention flow in the video recommendation network. We propose a model that accounts for the network effects for predicting video popularity, and we show it consistently outperforms the baselines. This model also identifies a group of artists gaining attention because of the recommendation network. Altogether, our observations and our models provide a new set of tools to better understand the impacts of recommender systems on collective social attention. |
Tasks | Recommendation Systems |
Published | 2019-08-20 |
URL | https://arxiv.org/abs/1908.07123v3 |
https://arxiv.org/pdf/1908.07123v3.pdf | |
PWC | https://paperswithcode.com/paper/estimating-attention-flow-in-online-video |
Repo | https://github.com/avalanchesiqi/networked-popularity |
Framework | none |