Paper Group AWR 204
Improving Exploration in Evolution Strategies for Deep Reinforcement Learning via a Population of Novelty-Seeking Agents. Towards Real-Time Advancement of Underwater Visual Quality with GAN. PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning. Bidirectional deep-readout echo state networks. PWLS-ULTRA: An Efficient Clustering an …
Improving Exploration in Evolution Strategies for Deep Reinforcement Learning via a Population of Novelty-Seeking Agents
Title | Improving Exploration in Evolution Strategies for Deep Reinforcement Learning via a Population of Novelty-Seeking Agents |
Authors | Edoardo Conti, Vashisht Madhavan, Felipe Petroski Such, Joel Lehman, Kenneth O. Stanley, Jeff Clune |
Abstract | Evolution strategies (ES) are a family of black-box optimization algorithms able to train deep neural networks roughly as well as Q-learning and policy gradient methods on challenging deep reinforcement learning (RL) problems, but are much faster (e.g. hours vs. days) because they parallelize better. However, many RL problems require directed exploration because they have reward functions that are sparse or deceptive (i.e. contain local optima), and it is unknown how to encourage such exploration with ES. Here we show that algorithms that have been invented to promote directed exploration in small-scale evolved neural networks via populations of exploring agents, specifically novelty search (NS) and quality diversity (QD) algorithms, can be hybridized with ES to improve its performance on sparse or deceptive deep RL tasks, while retaining scalability. Our experiments confirm that the resultant new algorithms, NS-ES and two QD algorithms, NSR-ES and NSRA-ES, avoid local optima encountered by ES to achieve higher performance on Atari and simulated robots learning to walk around a deceptive trap. This paper thus introduces a family of fast, scalable algorithms for reinforcement learning that are capable of directed exploration. It also adds this new family of exploration algorithms to the RL toolbox and raises the interesting possibility that analogous algorithms with multiple simultaneous paths of exploration might also combine well with existing RL algorithms outside ES. |
Tasks | Policy Gradient Methods, Q-Learning |
Published | 2017-12-18 |
URL | http://arxiv.org/abs/1712.06560v3 |
http://arxiv.org/pdf/1712.06560v3.pdf | |
PWC | https://paperswithcode.com/paper/improving-exploration-in-evolution-strategies |
Repo | https://github.com/uber-common/deep-neuroevolution |
Framework | tf |
Towards Real-Time Advancement of Underwater Visual Quality with GAN
Title | Towards Real-Time Advancement of Underwater Visual Quality with GAN |
Authors | Xingyu Chen, Junzhi Yu, Shihan Kong, Zhengxing Wu, Xi Fang, Li Wen |
Abstract | Low visual quality has prevented underwater robotic vision from a wide range of applications. Although several algorithms have been developed, real-time and adaptive methods are deficient for real-world tasks. In this paper, we address this difficulty based on generative adversarial networks (GAN), and propose a GAN-based restoration scheme (GAN-RS). In particular, we develop a multi-branch discriminator including an adversarial branch and a critic branch for the purpose of simultaneously preserving image content and removing underwater noise. In addition to adversarial learning, a novel dark channel prior loss also promotes the generator to produce realistic vision. More specifically, an underwater index is investigated to describe underwater properties, and a loss function based on the underwater index is designed to train the critic branch for underwater noise suppression. Through extensive comparisons on visual quality and feature restoration, we confirm the superiority of the proposed approach. Consequently, the GAN-RS can adaptively improve underwater visual quality in real time and induce an overall superior restoration performance. Finally, a real-world experiment is conducted on the seabed for grasping marine products, and the results are quite promising. The source code is publicly available at https://github.com/SeanChenxy/GAN_RS. |
Tasks | |
Published | 2017-12-03 |
URL | https://arxiv.org/abs/1712.00736v4 |
https://arxiv.org/pdf/1712.00736v4.pdf | |
PWC | https://paperswithcode.com/paper/towards-quality-advancement-of-underwater |
Repo | https://github.com/SeanChenxy/GAN_RS |
Framework | pytorch |
PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning
Title | PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning |
Authors | Arun Mallya, Svetlana Lazebnik |
Abstract | This paper presents a method for adding multiple tasks to a single deep neural network while avoiding catastrophic forgetting. Inspired by network pruning techniques, we exploit redundancies in large deep networks to free up parameters that can then be employed to learn new tasks. By performing iterative pruning and network re-training, we are able to sequentially “pack” multiple tasks into a single network while ensuring minimal drop in performance and minimal storage overhead. Unlike prior work that uses proxy losses to maintain accuracy on older tasks, we always optimize for the task at hand. We perform extensive experiments on a variety of network architectures and large-scale datasets, and observe much better robustness against catastrophic forgetting than prior work. In particular, we are able to add three fine-grained classification tasks to a single ImageNet-trained VGG-16 network and achieve accuracies close to those of separately trained networks for each task. Code available at https://github.com/arunmallya/packnet |
Tasks | Network Pruning |
Published | 2017-11-15 |
URL | http://arxiv.org/abs/1711.05769v2 |
http://arxiv.org/pdf/1711.05769v2.pdf | |
PWC | https://paperswithcode.com/paper/packnet-adding-multiple-tasks-to-a-single |
Repo | https://github.com/arunmallya/packnet |
Framework | pytorch |
Bidirectional deep-readout echo state networks
Title | Bidirectional deep-readout echo state networks |
Authors | Filippo Maria Bianchi, Simone Scardapane, Sigurd Løkse, Robert Jenssen |
Abstract | We propose a deep architecture for the classification of multivariate time series. By means of a recurrent and untrained reservoir we generate a vectorial representation that embeds temporal relationships in the data. To improve the memorization capability, we implement a bidirectional reservoir, whose last state captures also past dependencies in the input. We apply dimensionality reduction to the final reservoir states to obtain compressed fixed size representations of the time series. These are subsequently fed into a deep feedforward network trained to perform the final classification. We test our architecture on benchmark datasets and on a real-world use-case of blood samples classification. Results show that our method performs better than a standard echo state network and, at the same time, achieves results comparable to a fully-trained recurrent network, but with a faster training. |
Tasks | Dimensionality Reduction, Time Series |
Published | 2017-11-17 |
URL | http://arxiv.org/abs/1711.06509v3 |
http://arxiv.org/pdf/1711.06509v3.pdf | |
PWC | https://paperswithcode.com/paper/bidirectional-deep-readout-echo-state |
Repo | https://github.com/FilippoMB/Bidirectional-Deep-Echo-State-Network |
Framework | tf |
PWLS-ULTRA: An Efficient Clustering and Learning-Based Approach for Low-Dose 3D CT Image Reconstruction
Title | PWLS-ULTRA: An Efficient Clustering and Learning-Based Approach for Low-Dose 3D CT Image Reconstruction |
Authors | Xuehang Zheng, Saiprasad Ravishankar, Yong Long, Jeffrey A. Fessler |
Abstract | The development of computed tomography (CT) image reconstruction methods that significantly reduce patient radiation exposure while maintaining high image quality is an important area of research in low-dose CT (LDCT) imaging. We propose a new penalized weighted least squares (PWLS) reconstruction method that exploits regularization based on an efficient Union of Learned TRAnsforms (PWLS-ULTRA). The union of square transforms is pre-learned from numerous image patches extracted from a dataset of CT images or volumes. The proposed PWLS-based cost function is optimized by alternating between a CT image reconstruction step, and a sparse coding and clustering step. The CT image reconstruction step is accelerated by a relaxed linearized augmented Lagrangian method with ordered-subsets that reduces the number of forward and back projections. Simulations with 2-D and 3-D axial CT scans of the extended cardiac-torso phantom and 3D helical chest and abdomen scans show that for both normal-dose and low-dose levels, the proposed method significantly improves the quality of reconstructed images compared to PWLS reconstruction with a nonadaptive edge-preserving regularizer (PWLS-EP). PWLS with regularization based on a union of learned transforms leads to better image reconstructions than using a single learned square transform. We also incorporate patch-based weights in PWLS-ULTRA that enhance image quality and help improve image resolution uniformity. The proposed approach achieves comparable or better image quality compared to learned overcomplete synthesis dictionaries, but importantly, is much faster (computationally more efficient). |
Tasks | Computed Tomography (CT), Image Reconstruction |
Published | 2017-03-27 |
URL | http://arxiv.org/abs/1703.09165v3 |
http://arxiv.org/pdf/1703.09165v3.pdf | |
PWC | https://paperswithcode.com/paper/pwls-ultra-an-efficient-clustering-and |
Repo | https://github.com/xuehangzheng/PWLS-ULTRA-for-Low-Dose-3D-CT-Image-Reconstruction |
Framework | none |
Synthesis of Positron Emission Tomography (PET) Images via Multi-channel Generative Adversarial Networks (GANs)
Title | Synthesis of Positron Emission Tomography (PET) Images via Multi-channel Generative Adversarial Networks (GANs) |
Authors | Lei Bi, Jinman Kim, Ashnil Kumar, Dagan Feng, Michael Fulham |
Abstract | Positron emission tomography (PET) image synthesis plays an important role, which can be used to boost the training data for computer aided diagnosis systems. However, existing image synthesis methods have problems in synthesizing the low resolution PET images. To address these limitations, we propose multi-channel generative adversarial networks (M-GAN) based PET image synthesis method. Different to the existing methods which rely on using low-level features, the proposed M-GAN is capable to represent the features in a high-level of semantic based on the adversarial learning concept. In addition, M-GAN enables to take the input from the annotation (label) to synthesize the high uptake regions e.g., tumors and from the computed tomography (CT) images to constrain the appearance consistency and output the synthetic PET images directly. Our results on 50 lung cancer PET-CT studies indicate that our method was much closer to the real PET images when compared with the existing methods. |
Tasks | Computed Tomography (CT), Image Generation |
Published | 2017-07-31 |
URL | http://arxiv.org/abs/1707.09747v1 |
http://arxiv.org/pdf/1707.09747v1.pdf | |
PWC | https://paperswithcode.com/paper/synthesis-of-positron-emission-tomography-pet |
Repo | https://github.com/ChengBinJin/SpineC2M |
Framework | tf |
Measuring the Reliability of Hate Speech Annotations: The Case of the European Refugee Crisis
Title | Measuring the Reliability of Hate Speech Annotations: The Case of the European Refugee Crisis |
Authors | Björn Ross, Michael Rist, Guillermo Carbonell, Benjamin Cabrera, Nils Kurowsky, Michael Wojatzki |
Abstract | Some users of social media are spreading racist, sexist, and otherwise hateful content. For the purpose of training a hate speech detection system, the reliability of the annotations is crucial, but there is no universally agreed-upon definition. We collected potentially hateful messages and asked two groups of internet users to determine whether they were hate speech or not, whether they should be banned or not and to rate their degree of offensiveness. One of the groups was shown a definition prior to completing the survey. We aimed to assess whether hate speech can be annotated reliably, and the extent to which existing definitions are in accordance with subjective ratings. Our results indicate that showing users a definition caused them to partially align their own opinion with the definition but did not improve reliability, which was very low overall. We conclude that the presence of hate speech should perhaps not be considered a binary yes-or-no decision, and raters need more detailed instructions for the annotation. |
Tasks | Hate Speech Detection |
Published | 2017-01-27 |
URL | http://arxiv.org/abs/1701.08118v1 |
http://arxiv.org/pdf/1701.08118v1.pdf | |
PWC | https://paperswithcode.com/paper/measuring-the-reliability-of-hate-speech |
Repo | https://github.com/UCSM-DUE/IWG_hatespeech_public |
Framework | none |
Lucid Data Dreaming for Video Object Segmentation
Title | Lucid Data Dreaming for Video Object Segmentation |
Authors | Anna Khoreva, Rodrigo Benenson, Eddy Ilg, Thomas Brox, Bernt Schiele |
Abstract | Convolutional networks reach top quality in pixel-level video object segmentation but require a large amount of training data (1k~100k) to deliver such results. We propose a new training strategy which achieves state-of-the-art results across three evaluation datasets while using 20x~1000x less annotated data than competing methods. Our approach is suitable for both single and multiple object segmentation. Instead of using large training sets hoping to generalize across domains, we generate in-domain training data using the provided annotation on the first frame of each video to synthesize (“lucid dream”) plausible future video frames. In-domain per-video training data allows us to train high quality appearance- and motion-based models, as well as tune the post-processing stage. This approach allows to reach competitive results even when training from only a single annotated frame, without ImageNet pre-training. Our results indicate that using a larger training set is not automatically better, and that for the video object segmentation task a smaller training set that is closer to the target domain is more effective. This changes the mindset regarding how many training samples and general “objectness” knowledge are required for the video object segmentation task. |
Tasks | Multiple Object Tracking, Object Tracking, Semantic Segmentation, Video Object Segmentation, Video Semantic Segmentation |
Published | 2017-03-28 |
URL | http://arxiv.org/abs/1703.09554v5 |
http://arxiv.org/pdf/1703.09554v5.pdf | |
PWC | https://paperswithcode.com/paper/lucid-data-dreaming-for-video-object |
Repo | https://github.com/yelantingfeng/pyLucid |
Framework | none |
DSSD : Deconvolutional Single Shot Detector
Title | DSSD : Deconvolutional Single Shot Detector |
Authors | Cheng-Yang Fu, Wei Liu, Ananth Ranga, Ambrish Tyagi, Alexander C. Berg |
Abstract | The main contribution of this paper is an approach for introducing additional context into state-of-the-art general object detection. To achieve this we first combine a state-of-the-art classifier (Residual-101[14]) with a fast detection framework (SSD[18]). We then augment SSD+Residual-101 with deconvolution layers to introduce additional large-scale context in object detection and improve accuracy, especially for small objects, calling our resulting system DSSD for deconvolutional single shot detector. While these two contributions are easily described at a high-level, a naive implementation does not succeed. Instead we show that carefully adding additional stages of learned transformations, specifically a module for feed-forward connections in deconvolution and a new output module, enables this new approach and forms a potential way forward for further detection research. Results are shown on both PASCAL VOC and COCO detection. Our DSSD with $513 \times 513$ input achieves 81.5% mAP on VOC2007 test, 80.0% mAP on VOC2012 test, and 33.2% mAP on COCO, outperforming a state-of-the-art method R-FCN[3] on each dataset. |
Tasks | Object Detection |
Published | 2017-01-23 |
URL | http://arxiv.org/abs/1701.06659v1 |
http://arxiv.org/pdf/1701.06659v1.pdf | |
PWC | https://paperswithcode.com/paper/dssd-deconvolutional-single-shot-detector |
Repo | https://github.com/MTCloudVision/mxnet-dssd |
Framework | mxnet |
CSI: A Hybrid Deep Model for Fake News Detection
Title | CSI: A Hybrid Deep Model for Fake News Detection |
Authors | Natali Ruchansky, Sungyong Seo, Yan Liu |
Abstract | The topic of fake news has drawn attention both from the public and the academic communities. Such misinformation has the potential of affecting public opinion, providing an opportunity for malicious parties to manipulate the outcomes of public events such as elections. Because such high stakes are at play, automatically detecting fake news is an important, yet challenging problem that is not yet well understood. Nevertheless, there are three generally agreed upon characteristics of fake news: the text of an article, the user response it receives, and the source users promoting it. Existing work has largely focused on tailoring solutions to one particular characteristic which has limited their success and generality. In this work, we propose a model that combines all three characteristics for a more accurate and automated prediction. Specifically, we incorporate the behavior of both parties, users and articles, and the group behavior of users who propagate fake news. Motivated by the three characteristics, we propose a model called CSI which is composed of three modules: Capture, Score, and Integrate. The first module is based on the response and text; it uses a Recurrent Neural Network to capture the temporal pattern of user activity on a given article. The second module learns the source characteristic based on the behavior of users, and the two are integrated with the third module to classify an article as fake or not. Experimental analysis on real-world data demonstrates that CSI achieves higher accuracy than existing models, and extracts meaningful latent representations of both users and articles. |
Tasks | Fake News Detection |
Published | 2017-03-20 |
URL | http://arxiv.org/abs/1703.06959v4 |
http://arxiv.org/pdf/1703.06959v4.pdf | |
PWC | https://paperswithcode.com/paper/csi-a-hybrid-deep-model-for-fake-news |
Repo | https://github.com/soorism/CSI-Code |
Framework | none |
Ensemble Sales Forecasting Study in Semiconductor Industry
Title | Ensemble Sales Forecasting Study in Semiconductor Industry |
Authors | Qiuping Xu, Vikas Sharma |
Abstract | Sales forecasting plays a prominent role in business planning and business strategy. The value and importance of advance information is a cornerstone of planning activity, and a well-set forecast goal can guide sale-force more efficiently. In this paper CPU sales forecasting of Intel Corporation, a multinational semiconductor industry, was considered. Past sale, future booking, exchange rates, Gross domestic product (GDP) forecasting, seasonality and other indicators were innovatively incorporated into the quantitative modeling. Benefit from the recent advances in computation power and software development, millions of models built upon multiple regressions, time series analysis, random forest and boosting tree were executed in parallel. The models with smaller validation errors were selected to form the ensemble model. To better capture the distinct characteristics, forecasting models were implemented at lead time and lines of business level. The moving windows validation process automatically selected the models which closely represent current market condition. The weekly cadence forecasting schema allowed the model to response effectively to market fluctuation. Generic variable importance analysis was also developed to increase the model interpretability. Rather than assuming fixed distribution, this non-parametric permutation variable importance analysis provided a general framework across methods to evaluate the variable importance. This variable importance framework can further extend to classification problem by modifying the mean absolute percentage error(MAPE) into misclassify error. Please find the demo code at : https://github.com/qx0731/ensemble_forecast_methods |
Tasks | Time Series, Time Series Analysis |
Published | 2017-04-28 |
URL | http://arxiv.org/abs/1705.00003v3 |
http://arxiv.org/pdf/1705.00003v3.pdf | |
PWC | https://paperswithcode.com/paper/ensemble-sales-forecasting-study-in |
Repo | https://github.com/qx0731/ensemble_forecast_methods |
Framework | none |
(Quasi)Periodicity Quantification in Video Data, Using Topology
Title | (Quasi)Periodicity Quantification in Video Data, Using Topology |
Authors | Christopher J. Tralie, Jose A. Perea |
Abstract | This work introduces a novel framework for quantifying the presence and strength of recurrent dynamics in video data. Specifically, we provide continuous measures of periodicity (perfect repetition) and quasiperiodicity (superposition of periodic modes with non-commensurate periods), in a way which does not require segmentation, training, object tracking or 1-dimensional surrogate signals. Our methodology operates directly on video data. The approach combines ideas from nonlinear time series analysis (delay embeddings) and computational topology (persistent homology), by translating the problem of finding recurrent dynamics in video data, into the problem of determining the circularity or toroidality of an associated geometric space. Through extensive testing, we show the robustness of our scores with respect to several noise models/levels, we show that our periodicity score is superior to other methods when compared to human-generated periodicity rankings, and furthermore, we show that our quasiperiodicity score clearly indicates the presence of biphonation in videos of vibrating vocal folds, which has never before been accomplished end to end quantitatively. |
Tasks | Object Tracking, Time Series, Time Series Analysis |
Published | 2017-04-26 |
URL | http://arxiv.org/abs/1704.08382v2 |
http://arxiv.org/pdf/1704.08382v2.pdf | |
PWC | https://paperswithcode.com/paper/quasiperiodicity-quantification-in-video-data |
Repo | https://github.com/ctralie/SlidingWindowVideoTDA |
Framework | none |