October 20, 2019

2866 words 14 mins read

Paper Group AWR 187

Paper Group AWR 187

Cross-layer Optimization for High Speed Adders: A Pareto Driven Machine Learning Approach. B-DCGAN:Evaluation of Binarized DCGAN for FPGA. Automatic Classification of Music Genre using Masked Conditional Neural Networks. Deep Interest Evolution Network for Click-Through Rate Prediction. Gaussian Process Classification for Variable Fidelity Data. Mo …

Cross-layer Optimization for High Speed Adders: A Pareto Driven Machine Learning Approach

Title Cross-layer Optimization for High Speed Adders: A Pareto Driven Machine Learning Approach
Authors Yuzhe Ma, Subhendu Roy, Jin Miao, Jiamin Chen, Bei Yu
Abstract In spite of maturity to the modern electronic design automation (EDA) tools, optimized designs at architectural stage may become sub-optimal after going through physical design flow. Adder design has been such a long studied fundamental problem in VLSI industry yet designers cannot achieve optimal solutions by running EDA tools on the set of available prefix adder architectures. In this paper, we enhance a state-of-the-art prefix adder synthesis algorithm to obtain a much wider solution space in architectural domain. On top of that, a machine learning-based design space exploration methodology is applied to predict the Pareto frontier of the adders in physical domain, which is infeasible by exhaustively running EDA tools for innumerable architectural solutions. Considering the high cost of obtaining the true values for learning, an active learning algorithm is utilized to select the representative data during learning process, which uses less labeled data while achieving better quality of Pareto frontier. Experimental results demonstrate that our framework can achieve Pareto frontier of high quality over a wide design space, bridging the gap between architectural and physical designs.
Tasks Active Learning
Published 2018-07-18
URL http://arxiv.org/abs/1807.07023v2
PDF http://arxiv.org/pdf/1807.07023v2.pdf
PWC https://paperswithcode.com/paper/cross-layer-optimization-for-high-speed
Repo https://github.com/yuzhe630/adder-DSE
Framework none

B-DCGAN:Evaluation of Binarized DCGAN for FPGA

Title B-DCGAN:Evaluation of Binarized DCGAN for FPGA
Authors Hideo Terada, Hayaru Shouno
Abstract We are trying to implement deep neural networks in the edge computing environment for real-world applications such as the IoT(Internet of Things), the FinTech etc., for the purpose of utilizing the significant achievement of Deep Learning in recent years. Especially, we now focus algorithm implementation on FPGA, because FPGA is one of the promising devices for low-cost and low-power implementation of the edge computer. In this work, we introduce Binary-DCGAN(B-DCGAN) - Deep Convolutional GAN model with binary weights and activations, and with using integer-valued operations in forward pass(train-time and run-time). And we show how to implement B-DCGAN on FPGA(Xilinx Zynq). Using the B-DCGAN, we do feasibility study of FPGA’s characteristic and performance for Deep Learning. Because the binarization and using integer-valued operation reduce the memory capacity and the number of the circuit gates, it is very effective for FPGA implementation. On the other hand, the quality of generated data from the model will be decreased by these reductions. So we investigate the influence of these reductions.
Tasks
Published 2018-03-29
URL https://arxiv.org/abs/1803.10930v2
PDF https://arxiv.org/pdf/1803.10930v2.pdf
PWC https://paperswithcode.com/paper/b-dcganevaluation-of-binarized-dcgan-for-fpga
Repo https://github.com/hterada/b-dcgan
Framework none

Automatic Classification of Music Genre using Masked Conditional Neural Networks

Title Automatic Classification of Music Genre using Masked Conditional Neural Networks
Authors Fady Medhat, David Chesmore, John Robinson
Abstract Neural network based architectures used for sound recognition are usually adapted from other application domains such as image recognition, which may not harness the time-frequency representation of a signal. The ConditionaL Neural Networks (CLNN) and its extension the Masked ConditionaL Neural Networks (MCLNN) are designed for multidimensional temporal signal recognition. The CLNN is trained over a window of frames to preserve the inter-frame relation, and the MCLNN enforces a systematic sparseness over the network’s links that mimics a filterbank-like behavior. The masking operation induces the network to learn in frequency bands, which decreases the network susceptibility to frequency-shifts in time-frequency representations. Additionally, the mask allows an exploration of a range of feature combinations concurrently analogous to the manual handcrafting of the optimum collection of features for a recognition task. MCLNN have achieved competitive performance on the Ballroom music dataset compared to several hand-crafted attempts and outperformed models based on state-of-the-art Convolutional Neural Networks.
Tasks
Published 2018-01-16
URL http://arxiv.org/abs/1801.05504v2
PDF http://arxiv.org/pdf/1801.05504v2.pdf
PWC https://paperswithcode.com/paper/automatic-classification-of-music-genre-using
Repo https://github.com/fadymedhat/MCLNN
Framework tf

Deep Interest Evolution Network for Click-Through Rate Prediction

Title Deep Interest Evolution Network for Click-Through Rate Prediction
Authors Guorui Zhou, Na Mou, Ying Fan, Qi Pi, Weijie Bian, Chang Zhou, Xiaoqiang Zhu, Kun Gai
Abstract Easy-to-use,Modular and Extendible package of deep-learning based CTR models.DeepFM,DeepInterestNetwork(DIN),DeepInterestEvolutionNetwork(DIEN),DeepCrossNetwork(DCN),AttentionalFactorizationMachine(AFM),Neural Factorization Machine(NFM),AutoInt
Tasks Click-Through Rate Prediction
Published 2018-09-11
URL http://arxiv.org/abs/1809.03672v5
PDF http://arxiv.org/pdf/1809.03672v5.pdf
PWC https://paperswithcode.com/paper/deep-interest-evolution-network-for-click
Repo https://github.com/mouna99/dien
Framework tf

Gaussian Process Classification for Variable Fidelity Data

Title Gaussian Process Classification for Variable Fidelity Data
Authors Nikita Klyuchnikov, Evgeny Burnaev
Abstract In this paper we address a classification problem where two sources of labels with different levels of fidelity are available. Our approach is to combine data from both sources by applying a co-kriging schema on latent functions, which allows the model to account item-dependent labeling discrepancy. We provide an extension of Laplace inference for Gaussian process classification, that takes into account multi-fidelity data. We evaluate the proposed method on real and synthetic datasets and show that it is more resistant to different levels of discrepancy between sources than other approaches for data fusion. Our method can provide accuracy/cost trade-off for a number of practical tasks such as crowd-sourced data annotation and feasibility regions construction in engineering design.
Tasks
Published 2018-09-13
URL https://arxiv.org/abs/1809.05143v3
PDF https://arxiv.org/pdf/1809.05143v3.pdf
PWC https://paperswithcode.com/paper/gaussian-process-classification-for-variable
Repo https://github.com/user525/mfgpc
Framework none

Moments in Time Dataset: one million videos for event understanding

Title Moments in Time Dataset: one million videos for event understanding
Authors Mathew Monfort, Alex Andonian, Bolei Zhou, Kandan Ramakrishnan, Sarah Adel Bargal, Tom Yan, Lisa Brown, Quanfu Fan, Dan Gutfruend, Carl Vondrick, Aude Oliva
Abstract We present the Moments in Time Dataset, a large-scale human-annotated collection of one million short videos corresponding to dynamic events unfolding within three seconds. Modeling the spatial-audio-temporal dynamics even for actions occurring in 3 second videos poses many challenges: meaningful events do not include only people, but also objects, animals, and natural phenomena; visual and auditory events can be symmetrical in time (“opening” is “closing” in reverse), and either transient or sustained. We describe the annotation process of our dataset (each video is tagged with one action or activity label among 339 different classes), analyze its scale and diversity in comparison to other large-scale video datasets for action recognition, and report results of several baseline models addressing separately, and jointly, three modalities: spatial, temporal and auditory. The Moments in Time dataset, designed to have a large coverage and diversity of events in both visual and auditory modalities, can serve as a new challenge to develop models that scale to the level of complexity and abstract reasoning that a human processes on a daily basis.
Tasks Action Recognition In Videos, Multimodal Activity Recognition, Temporal Action Localization
Published 2018-01-09
URL http://arxiv.org/abs/1801.03150v3
PDF http://arxiv.org/pdf/1801.03150v3.pdf
PWC https://paperswithcode.com/paper/moments-in-time-dataset-one-million-videos
Repo https://github.com/shubhambitsg/activity-recognition
Framework pytorch

Competitive Inner-Imaging Squeeze and Excitation for Residual Network

Title Competitive Inner-Imaging Squeeze and Excitation for Residual Network
Authors Yang Hu, Guihua Wen, Mingnan Luo, Dan Dai, Jiajiong Ma, Zhiwen Yu
Abstract Residual networks, which use a residual unit to supplement the identity mappings, enable very deep convolutional architecture to operate well, however, the residual architecture has been proved to be diverse and redundant, which may leads to low-efficient modeling. In this work, we propose a competitive squeeze-excitation (SE) mechanism for the residual network. Re-scaling the value for each channel in this structure will be determined by the residual and identity mappings jointly, and this design enables us to expand the meaning of channel relationship modeling in residual blocks. Modeling of the competition between residual and identity mappings cause the identity flow to control the complement of the residual feature maps for itself. Furthermore, we design a novel inner-imaging competitive SE block to shrink the consumption and re-image the global features of intermediate network structure, by using the inner-imaging mechanism, we can model the channel-wise relations with convolution in spatial. We carry out experiments on the CIFAR, SVHN, and ImageNet datasets, and the proposed method can challenge state-of-the-art results.
Tasks
Published 2018-07-24
URL http://arxiv.org/abs/1807.08920v4
PDF http://arxiv.org/pdf/1807.08920v4.pdf
PWC https://paperswithcode.com/paper/competitive-inner-imaging-squeeze-and
Repo https://github.com/scut-aitcm/Competitive-Inner-Imaging-SENet
Framework tf

Factorized Machine Self-Confidence for Decision-Making Agents

Title Factorized Machine Self-Confidence for Decision-Making Agents
Authors Brett W Israelsen, Nisar R Ahmed, Eric Frew, Dale Lawrence, Brian Argrow
Abstract Algorithmic assurances from advanced autonomous systems assist human users in understanding, trusting, and using such systems appropriately. Designing these systems with the capacity of assessing their own capabilities is one approach to creating an algorithmic assurance. The idea of machine self-confidence' is introduced for autonomous systems. Using a factorization based framework for self-confidence assessment, one component of self-confidence, called solver-quality’, is discussed in the context of Markov decision processes for autonomous systems. Markov decision processes underlie much of the theory of reinforcement learning, and are commonly used for planning and decision making under uncertainty in robotics and autonomous systems. A `solver quality’ metric is formally defined in the context of decision making algorithms based on Markov decision processes. A method for assessing solver quality is then derived, drawing inspiration from empirical hardness models. Finally, numerical experiments for an unmanned autonomous vehicle navigation problem under different solver, parameter, and environment conditions indicate that the self-confidence metric exhibits the desired properties. Discussion of results, and avenues for future investigation are included. |
Tasks Decision Making, Decision Making Under Uncertainty
Published 2018-10-15
URL http://arxiv.org/abs/1810.06519v2
PDF http://arxiv.org/pdf/1810.06519v2.pdf
PWC https://paperswithcode.com/paper/factorized-machine-self-confidence-for
Repo https://github.com/COHRINT/FaMSeC
Framework mxnet

A Frank-Wolfe Framework for Efficient and Effective Adversarial Attacks

Title A Frank-Wolfe Framework for Efficient and Effective Adversarial Attacks
Authors Jinghui Chen, Dongruo Zhou, Jinfeng Yi, Quanquan Gu
Abstract Depending on how much information an adversary can access to, adversarial attacks can be classified as white-box attack and black-box attack. For white-box attack, optimization-based attack algorithms such as projected gradient descent (PGD) can achieve relatively high attack success rates within moderate iterates. However, they tend to generate adversarial examples near or upon the boundary of the perturbation set, resulting in large distortion. Furthermore, their corresponding black-box attack algorithms also suffer from high query complexities, thereby limiting their practical usefulness. In this paper, we focus on the problem of developing efficient and effective optimization-based adversarial attack algorithms. In particular, we propose a novel adversarial attack framework for both white-box and black-box settings based on a variant of Frank-Wolfe algorithm. We show in theory that the proposed attack algorithms are efficient with an $O(1/\sqrt{T})$ convergence rate. The empirical results of attacking the ImageNet and MNIST datasets also verify the efficiency and effectiveness of the proposed algorithms. More specifically, our proposed algorithms attain the best attack performances in both white-box and black-box attacks among all baselines, and are more time and query efficient than the state-of-the-art.
Tasks Adversarial Attack
Published 2018-11-27
URL https://arxiv.org/abs/1811.10828v2
PDF https://arxiv.org/pdf/1811.10828v2.pdf
PWC https://paperswithcode.com/paper/a-frank-wolfe-framework-for-efficient-and
Repo https://github.com/sipka/fw-adversarial-imagenet
Framework none

High-Accuracy Low-Precision Training

Title High-Accuracy Low-Precision Training
Authors Christopher De Sa, Megan Leszczynski, Jian Zhang, Alana Marzoev, Christopher R. Aberger, Kunle Olukotun, Christopher Ré
Abstract Low-precision computation is often used to lower the time and energy cost of machine learning, and recently hardware accelerators have been developed to support it. Still, it has been used primarily for inference - not training. Previous low-precision training algorithms suffered from a fundamental tradeoff: as the number of bits of precision is lowered, quantization noise is added to the model, which limits statistical accuracy. To address this issue, we describe a simple low-precision stochastic gradient descent variant called HALP. HALP converges at the same theoretical rate as full-precision algorithms despite the noise introduced by using low precision throughout execution. The key idea is to use SVRG to reduce gradient variance, and to combine this with a novel technique called bit centering to reduce quantization error. We show that on the CPU, HALP can run up to $4 \times$ faster than full-precision SVRG and can match its convergence trajectory. We implemented HALP in TensorQuant, and show that it exceeds the validation performance of plain low-precision SGD on two deep learning tasks.
Tasks Quantization
Published 2018-03-09
URL http://arxiv.org/abs/1803.03383v1
PDF http://arxiv.org/pdf/1803.03383v1.pdf
PWC https://paperswithcode.com/paper/high-accuracy-low-precision-training
Repo https://github.com/HazyResearch/lp_rffs
Framework pytorch

Hierarchical Neural Networks for Sequential Sentence Classification in Medical Scientific Abstracts

Title Hierarchical Neural Networks for Sequential Sentence Classification in Medical Scientific Abstracts
Authors Di Jin, Peter Szolovits
Abstract Prevalent models based on artificial neural network (ANN) for sentence classification often classify sentences in isolation without considering the context in which sentences appear. This hampers the traditional sentence classification approaches to the problem of sequential sentence classification, where structured prediction is needed for better overall classification performance. In this work, we present a hierarchical sequential labeling network to make use of the contextual information within surrounding sentences to help classify the current sentence. Our model outperforms the state-of-the-art results by 2%-3% on two benchmarking datasets for sequential sentence classification in medical scientific abstracts.
Tasks Sentence Classification, Structured Prediction
Published 2018-08-19
URL http://arxiv.org/abs/1808.06161v1
PDF http://arxiv.org/pdf/1808.06161v1.pdf
PWC https://paperswithcode.com/paper/hierarchical-neural-networks-for-sequential
Repo https://github.com/jind11/HSLN-Joint-Sentence-Classification
Framework tf

Adversarial Attacks and Defences Competition

Title Adversarial Attacks and Defences Competition
Authors Alexey Kurakin, Ian Goodfellow, Samy Bengio, Yinpeng Dong, Fangzhou Liao, Ming Liang, Tianyu Pang, Jun Zhu, Xiaolin Hu, Cihang Xie, Jianyu Wang, Zhishuai Zhang, Zhou Ren, Alan Yuille, Sangxia Huang, Yao Zhao, Yuzhe Zhao, Zhonglin Han, Junjiajia Long, Yerkebulan Berdibekov, Takuya Akiba, Seiya Tokui, Motoki Abe
Abstract To accelerate research on adversarial examples and robustness of machine learning classifiers, Google Brain organized a NIPS 2017 competition that encouraged researchers to develop new methods to generate adversarial examples as well as to develop new ways to defend against them. In this chapter, we describe the structure and organization of the competition and the solutions developed by several of the top-placing teams.
Tasks
Published 2018-03-31
URL http://arxiv.org/abs/1804.00097v1
PDF http://arxiv.org/pdf/1804.00097v1.pdf
PWC https://paperswithcode.com/paper/adversarial-attacks-and-defences-competition
Repo https://github.com/pfnet-research/nips17-adversarial-attack
Framework tf

Sym-parameterized Dynamic Inference for Mixed-Domain Image Translation

Title Sym-parameterized Dynamic Inference for Mixed-Domain Image Translation
Authors Simyung Chang, SeongUk Park, John Yang, Nojun Kwak
Abstract Recent advances in image-to-image translation have led to some ways to generate multiple domain images through a single network. However, there is still a limit in creating an image of a target domain without a dataset on it. We propose a method that expands the concept of `multi-domain’ from data to the loss area and learns the combined characteristics of each domain to dynamically infer translations of images in mixed domains. First, we introduce Sym-parameter and its learning method for variously mixed losses while synchronizing them with input conditions. Then, we propose Sym-parameterized Generative Network (SGN) which is empirically confirmed of learning mixed characteristics of various data and losses, and translating images to any mixed-domain without ground truths, such as 30% Van Gogh and 20% Monet and 40% snowy. |
Tasks Image-to-Image Translation
Published 2018-11-29
URL https://arxiv.org/abs/1811.12362v3
PDF https://arxiv.org/pdf/1811.12362v3.pdf
PWC https://paperswithcode.com/paper/image-translation-to-mixed-domain-using-sym
Repo https://github.com/TimeLighter/pytorch-sym-parameter
Framework pytorch

Deep EHR: Chronic Disease Prediction Using Medical Notes

Title Deep EHR: Chronic Disease Prediction Using Medical Notes
Authors Jingshu Liu, Zachariah Zhang, Narges Razavian
Abstract Early detection of preventable diseases is important for better disease management, improved inter-ventions, and more efficient health-care resource allocation. Various machine learning approacheshave been developed to utilize information in Electronic Health Record (EHR) for this task. Majorityof previous attempts, however, focus on structured fields and lose the vast amount of information inthe unstructured notes. In this work we propose a general multi-task framework for disease onsetprediction that combines both free-text medical notes and structured information. We compareperformance of different deep learning architectures including CNN, LSTM and hierarchical models.In contrast to traditional text-based prediction models, our approach does not require disease specificfeature engineering, and can handle negations and numerical values that exist in the text. Ourresults on a cohort of about 1 million patients show that models using text outperform modelsusing just structured data, and that models capable of using numerical values and negations in thetext, in addition to the raw text, further improve performance. Additionally, we compare differentvisualization methods for medical professionals to interpret model predictions.
Tasks Disease Prediction
Published 2018-08-15
URL http://arxiv.org/abs/1808.04928v1
PDF http://arxiv.org/pdf/1808.04928v1.pdf
PWC https://paperswithcode.com/paper/deep-ehr-chronic-disease-prediction-using
Repo https://github.com/NYUMedML/DeepEHR
Framework pytorch

TOM-Net: Learning Transparent Object Matting from a Single Image

Title TOM-Net: Learning Transparent Object Matting from a Single Image
Authors Guanying Chen, Kai Han, Kwan-Yee K. Wong
Abstract This paper addresses the problem of transparent object matting. Existing image matting approaches for transparent objects often require tedious capturing procedures and long processing time, which limit their practical use. In this paper, we first formulate transparent object matting as a refractive flow estimation problem. We then propose a deep learning framework, called TOM-Net, for learning the refractive flow. Our framework comprises two parts, namely a multi-scale encoder-decoder network for producing a coarse prediction, and a residual network for refinement. At test time, TOM-Net takes a single image as input, and outputs a matte (consisting of an object mask, an attenuation mask and a refractive flow field) in a fast feed-forward pass. As no off-the-shelf dataset is available for transparent object matting, we create a large-scale synthetic dataset consisting of 158K images of transparent objects rendered in front of images sampled from the Microsoft COCO dataset. We also collect a real dataset consisting of 876 samples using 14 transparent objects and 60 background images. Promising experimental results have been achieved on both synthetic and real data, which clearly demonstrate the effectiveness of our approach.
Tasks Image Matting
Published 2018-03-13
URL http://arxiv.org/abs/1803.04636v3
PDF http://arxiv.org/pdf/1803.04636v3.pdf
PWC https://paperswithcode.com/paper/tom-net-learning-transparent-object-matting
Repo https://github.com/guanyingc/TOM-Net
Framework pytorch
comments powered by Disqus