Paper Group AWR 456
FasterSeg: Searching for Faster Real-time Semantic Segmentation. BOBBY2: Buffer Based Robust High-Speed Object Tracking. Unsupervised Neural Quantization for Compressed-Domain Similarity Search. Revealing the Importance of Semantic Retrieval for Machine Reading at Scale. Random Pairwise Shapelets Forest. SurfCon: Synonym Discovery on Privacy-Aware …
FasterSeg: Searching for Faster Real-time Semantic Segmentation
Title | FasterSeg: Searching for Faster Real-time Semantic Segmentation |
Authors | Wuyang Chen, Xinyu Gong, Xianming Liu, Qian Zhang, Yuan Li, Zhangyang Wang |
Abstract | We present FasterSeg, an automatically designed semantic segmentation network with not only state-of-the-art performance but also faster speed than current methods. Utilizing neural architecture search (NAS), FasterSeg is discovered from a novel and broader search space integrating multi-resolution branches, that has been recently found to be vital in manually designed segmentation models. To better calibrate the balance between the goals of high accuracy and low latency, we propose a decoupled and fine-grained latency regularization, that effectively overcomes our observed phenomenons that the searched networks are prone to “collapsing” to low-latency yet poor-accuracy models. Moreover, we seamlessly extend FasterSeg to a new collaborative search (co-searching) framework, simultaneously searching for a teacher and a student network in the same single run. The teacher-student distillation further boosts the student model’s accuracy. Experiments on popular segmentation benchmarks demonstrate the competency of FasterSeg. For example, FasterSeg can run over 30% faster than the closest manually designed competitor on Cityscapes, while maintaining comparable accuracy. |
Tasks | Neural Architecture Search, Real-Time Semantic Segmentation, Semantic Segmentation |
Published | 2019-12-23 |
URL | https://arxiv.org/abs/1912.10917v2 |
https://arxiv.org/pdf/1912.10917v2.pdf | |
PWC | https://paperswithcode.com/paper/fasterseg-searching-for-faster-real-time-1 |
Repo | https://github.com/TAMU-VITA/FasterSeg |
Framework | pytorch |
BOBBY2: Buffer Based Robust High-Speed Object Tracking
Title | BOBBY2: Buffer Based Robust High-Speed Object Tracking |
Authors | Keifer Lee, Jun Jet Tai, Swee King Phang |
Abstract | In this work, a novel high-speed single object tracker that is robust against non-semantic distractor exemplars is introduced; dubbed BOBBY2. It incorporates a novel exemplar buffer module that sparsely caches the target’s appearance across time, enabling it to adapt to potential target deformation. As for training, an augmented ImageNet-VID dataset was used in conjunction with the one cycle policy, enabling it to reach convergence with less than 2 epoch worth of data. For validation, the model was benchmarked on the GOT-10k dataset and on an additional small, albeit challenging custom UAV dataset collected with the TU-3 UAV. We demonstrate that the exemplar buffer is capable of providing redundancies in case of unintended target drifts, a desirable trait in any middle to long term tracking. Even when the buffer is predominantly filled with distractors instead of valid exemplars, BOBBY2 is capable of maintaining a near-optimal level of accuracy. BOBBY2 manages to achieve a very competitive result on the GOT-10k dataset and to a lesser degree on the challenging custom TU-3 dataset, without fine-tuning, demonstrating its generalizability. In terms of speed, BOBBY2 utilizes a stripped down AlexNet as feature extractor with 63% less parameters than a vanilla AlexNet, thus being able to run at a competitive 85 FPS. |
Tasks | Object Tracking |
Published | 2019-10-18 |
URL | https://arxiv.org/abs/1910.08263v1 |
https://arxiv.org/pdf/1910.08263v1.pdf | |
PWC | https://paperswithcode.com/paper/bobby2-buffer-based-robust-high-speed-object |
Repo | https://github.com/datacrisis/BOBBY2 |
Framework | pytorch |
Unsupervised Neural Quantization for Compressed-Domain Similarity Search
Title | Unsupervised Neural Quantization for Compressed-Domain Similarity Search |
Authors | Stanislav Morozov, Artem Babenko |
Abstract | We tackle the problem of unsupervised visual descriptors compression, which is a key ingredient of large-scale image retrieval systems. While the deep learning machinery has benefited literally all computer vision pipelines, the existing state-of-the-art compression methods employ shallow architectures, and we aim to close this gap by our paper. In more detail, we introduce a DNN architecture for the unsupervised compressed-domain retrieval, based on multi-codebook quantization. The proposed architecture is designed to incorporate both fast data encoding and efficient distances computation via lookup tables. We demonstrate the exceptional advantage of our scheme over existing quantization approaches on several datasets of visual descriptors via outperforming the previous state-of-the-art by a large margin. |
Tasks | Image Retrieval, Quantization |
Published | 2019-08-11 |
URL | https://arxiv.org/abs/1908.03883v1 |
https://arxiv.org/pdf/1908.03883v1.pdf | |
PWC | https://paperswithcode.com/paper/unsupervised-neural-quantization-for |
Repo | https://github.com/stanis-morozov/unq |
Framework | pytorch |
Revealing the Importance of Semantic Retrieval for Machine Reading at Scale
Title | Revealing the Importance of Semantic Retrieval for Machine Reading at Scale |
Authors | Yixin Nie, Songhe Wang, Mohit Bansal |
Abstract | Machine Reading at Scale (MRS) is a challenging task in which a system is given an input query and is asked to produce a precise output by “reading” information from a large knowledge base. The task has gained popularity with its natural combination of information retrieval (IR) and machine comprehension (MC). Advancements in representation learning have led to separated progress in both IR and MC; however, very few studies have examined the relationship and combined design of retrieval and comprehension at different levels of granularity, for development of MRS systems. In this work, we give general guidelines on system design for MRS by proposing a simple yet effective pipeline system with special consideration on hierarchical semantic retrieval at both paragraph and sentence level, and their potential effects on the downstream task. The system is evaluated on both fact verification and open-domain multihop QA, achieving state-of-the-art results on the leaderboard test sets of both FEVER and HOTPOTQA. To further demonstrate the importance of semantic retrieval, we present ablation and analysis studies to quantify the contribution of neural retrieval modules at both paragraph-level and sentence-level, and illustrate that intermediate semantic retrieval modules are vital for not only effectively filtering upstream information and thus saving downstream computation, but also for shaping upstream data distribution and providing better data for downstream modeling. Code/data made publicly available at: https://github.com/easonnie/semanticRetrievalMRS |
Tasks | Information Retrieval, Reading Comprehension, Representation Learning |
Published | 2019-09-17 |
URL | https://arxiv.org/abs/1909.08041v1 |
https://arxiv.org/pdf/1909.08041v1.pdf | |
PWC | https://paperswithcode.com/paper/revealing-the-importance-of-semantic |
Repo | https://github.com/easonnie/semanticRetrievalMRS |
Framework | pytorch |
Random Pairwise Shapelets Forest
Title | Random Pairwise Shapelets Forest |
Authors | Mohan Shi, Zhihai Wang, Jodong Yuan, Haiyang Liu |
Abstract | Shapelet is a discriminative subsequence of time series. An advanced shapelet-based method is to embed shapelet into accurate and fast random forest. However, it shows several limitations. First, random shapelet forest requires a large training cost for split threshold searching. Second, a single shapelet provides limited information for only one branch of the decision tree, resulting in insufficient accuracy and interpretability. Third, randomized ensemble causes interpretability declining. For that, this paper presents Random Pairwise Shapelets Forest (RPSF). RPSF combines a pair of shapelets from different classes to construct random forest. It omits threshold searching to be more efficient, includes more information for each node of the forest to be more effective. Moreover, a discriminability metric, Decomposed Mean Decrease Impurity (DMDI), is proposed to identify influential region for every class. Extensive experiments show RPSF improves the accuracy and training speed of shapelet-based forest. Case studies demonstrate the interpretability of our method. |
Tasks | Time Series |
Published | 2019-03-19 |
URL | http://arxiv.org/abs/1903.07799v2 |
http://arxiv.org/pdf/1903.07799v2.pdf | |
PWC | https://paperswithcode.com/paper/random-pairwise-shapelets-forest |
Repo | https://github.com/nephashi/RandomPairwiseShapeletsForest |
Framework | none |
SurfCon: Synonym Discovery on Privacy-Aware Clinical Data
Title | SurfCon: Synonym Discovery on Privacy-Aware Clinical Data |
Authors | Zhen Wang, Xiang Yue, Soheil Moosavinasab, Yungui Huang, Simon Lin, Huan Sun |
Abstract | Unstructured clinical texts contain rich health-related information. To better utilize the knowledge buried in clinical texts, discovering synonyms for a medical query term has become an important task. Recent automatic synonym discovery methods leveraging raw text information have been developed. However, to preserve patient privacy and security, it is usually quite difficult to get access to large-scale raw clinical texts. In this paper, we study a new setting named synonym discovery on privacy-aware clinical data (i.e., medical terms extracted from the clinical texts and their aggregated co-occurrence counts, without raw clinical texts). To solve the problem, we propose a new framework SurfCon that leverages two important types of information in the privacy-aware clinical data, i.e., the surface form information, and the global context information for synonym discovery. In particular, the surface form module enables us to detect synonyms that look similar while the global context module plays a complementary role to discover synonyms that are semantically similar but in different surface forms, and both allow us to deal with the OOV query issue (i.e., when the query is not found in the given data). We conduct extensive experiments and case studies on publicly available privacy-aware clinical data, and show that SurfCon can outperform strong baseline methods by large margins under various settings. |
Tasks | |
Published | 2019-06-21 |
URL | https://arxiv.org/abs/1906.09285v1 |
https://arxiv.org/pdf/1906.09285v1.pdf | |
PWC | https://paperswithcode.com/paper/surfcon-synonym-discovery-on-privacy-aware |
Repo | https://github.com/yzabc007/SurfCon |
Framework | pytorch |
Representation Learning with Statistical Independence to Mitigate Bias
Title | Representation Learning with Statistical Independence to Mitigate Bias |
Authors | Ehsan Adeli, Qingyu Zhao, Adolf Pfefferbaum, Edith V. Sullivan, Li Fei-Fei, Juan Carlos Niebles, Kilian M. Pohl |
Abstract | Presence of bias (in datasets or tasks) is inarguably one of the most critical challenges in machine learning applications that has alluded to pivotal debates in recent years. Such challenges range from spurious associations between variables in medical studies to the bias of race in gender or face recognition systems. Controlling for all types of biases in the dataset curation stage is cumbersome and sometimes impossible. The alternative is to use the available data and build models incorporating fair representation learning. In this paper, we propose such a model based on adversarial training with two competing objectives to learn features that have (1) maximum discriminative power with respect to the task and (2) minimal statistical mean dependence with the protected (bias) variable(s). Our approach does so by incorporating a new adversarial loss function that encourages a vanished correlation between the bias and the learned features. We apply our method to synthetic data, medical images (containing task bias), and a dataset for gender classification (containing dataset bias). Our results show that the learned features by our method not only result in superior prediction performance but also are unbiased. The code is available at https://github.com/QingyuZhao/BR-Net/. |
Tasks | Face Recognition, Representation Learning |
Published | 2019-10-08 |
URL | https://arxiv.org/abs/1910.03676v2 |
https://arxiv.org/pdf/1910.03676v2.pdf | |
PWC | https://paperswithcode.com/paper/bias-resilient-neural-network |
Repo | https://github.com/QingyuZhao/BR-Net |
Framework | none |
AmazonQA: A Review-Based Question Answering Task
Title | AmazonQA: A Review-Based Question Answering Task |
Authors | Mansi Gupta, Nitish Kulkarni, Raghuveer Chanda, Anirudha Rayasam, Zachary C Lipton |
Abstract | Every day, thousands of customers post questions on Amazon product pages. After some time, if they are fortunate, a knowledgeable customer might answer their question. Observing that many questions can be answered based upon the available product reviews, we propose the task of review-based QA. Given a corpus of reviews and a question, the QA system synthesizes an answer. To this end, we introduce a new dataset and propose a method that combines information retrieval techniques for selecting relevant reviews (given a question) and “reading comprehension” models for synthesizing an answer (given a question and review). Our dataset consists of 923k questions, 3.6M answers and 14M reviews across 156k products. Building on the well-known Amazon dataset, we collect additional annotations, marking each question as either answerable or unanswerable based on the available reviews. A deployed system could first classify a question as answerable and then attempt to generate an answer. Notably, unlike many popular QA datasets, here, the questions, passages, and answers are all extracted from real human interactions. We evaluate numerous models for answer generation and propose strong baselines, demonstrating the challenging nature of this new task. |
Tasks | Information Retrieval, Question Answering, Reading Comprehension |
Published | 2019-08-12 |
URL | https://arxiv.org/abs/1908.04364v2 |
https://arxiv.org/pdf/1908.04364v2.pdf | |
PWC | https://paperswithcode.com/paper/amazonqa-a-review-based-question-answering |
Repo | https://github.com/amazonqa/amazonqa |
Framework | tf |
U-Net Fixed-Point Quantization for Medical Image Segmentation
Title | U-Net Fixed-Point Quantization for Medical Image Segmentation |
Authors | MohammadHossein AskariHemmat, Sina Honari, Lucas Rouhier, Christian S. Perone, Julien Cohen-Adad, Yvon Savaria, Jean-Pierre David |
Abstract | Model quantization is leveraged to reduce the memory consumption and the computation time of deep neural networks. This is achieved by representing weights and activations with a lower bit resolution when compared to their high precision floating point counterparts. The suitable level of quantization is directly related to the model performance. Lowering the quantization precision (e.g. 2 bits), reduces the amount of memory required to store model parameters and the amount of logic required to implement computational blocks, which contributes to reducing the power consumption of the entire system. These benefits typically come at the cost of reduced accuracy. The main challenge is to quantize a network as much as possible, while maintaining the performance accuracy. In this work, we present a quantization method for the U-Net architecture, a popular model in medical image segmentation. We then apply our quantization algorithm to three datasets: (1) the Spinal Cord Gray Matter Segmentation (GM), (2) the ISBI challenge for segmentation of neuronal structures in Electron Microscopic (EM), and (3) the public National Institute of Health (NIH) dataset for pancreas segmentation in abdominal CT scans. The reported results demonstrate that with only 4 bits for weights and 6 bits for activations, we obtain 8 fold reduction in memory requirements while loosing only 2.21%, 0.57% and 2.09% dice overlap score for EM, GM and NIH datasets respectively. Our fixed point quantization provides a flexible trade off between accuracy and memory requirement which is not provided by previous quantization methods for U-Net such as TernaryNet. |
Tasks | Medical Image Segmentation, Pancreas Segmentation, Quantization, Semantic Segmentation |
Published | 2019-08-02 |
URL | https://arxiv.org/abs/1908.01073v2 |
https://arxiv.org/pdf/1908.01073v2.pdf | |
PWC | https://paperswithcode.com/paper/u-net-fixed-point-quantization-for-medical |
Repo | https://github.com/paraficial/vae_pancreas_segmentation |
Framework | pytorch |
Towards Large yet Imperceptible Adversarial Image Perturbations with Perceptual Color Distance
Title | Towards Large yet Imperceptible Adversarial Image Perturbations with Perceptual Color Distance |
Authors | Zhengyu Zhao, Zhuoran Liu, Martha Larson |
Abstract | The success of image perturbations that are designed to fool image classifier is assessed in terms of both adversarial effect and visual imperceptibility. The conventional assumption on imperceptibility is that perturbations should strive for tight $L_p$-norm bounds in RGB space. In this work, we drop this assumption by pursuing an approach that exploits human color perception, and more specifically, minimizing perturbation size with respect to perceptual color distance. Our first approach, Perceptual Color distance C&W (PerC-C&W), extends the widely-used C&W approach and produces larger RGB perturbations. PerC-C&W is able to maintain adversarial strength, while contributing to imperceptibility. Our second approach, Perceptual Color distance Alternating Loss (PerC-AL), achieves the same outcome, but does so more efficiently by alternating between the classification loss and perceptual color difference when updating perturbations. Experimental evaluation shows PerC approaches outperform conventional $L_p$ approaches in terms of robustness and transferability, and also demonstrates that the PerC distance can provide added value on top of existing structure-based methods to creating image perturbations. |
Tasks | Image Classification |
Published | 2019-11-06 |
URL | https://arxiv.org/abs/1911.02466v2 |
https://arxiv.org/pdf/1911.02466v2.pdf | |
PWC | https://paperswithcode.com/paper/towards-large-yet-imperceptible-adversarial |
Repo | https://github.com/ZhengyuZhao/PerC-Adversarial |
Framework | pytorch |
Light Multi-segment Activation for Model Compression
Title | Light Multi-segment Activation for Model Compression |
Authors | Zhenhui Xu, Guolin Ke, Jia Zhang, Jiang Bian, Tie-Yan Liu |
Abstract | Model compression has become necessary when applying neural networks (NN) into many real application tasks that can accept slightly-reduced model accuracy with strict tolerance to model complexity. Recently, Knowledge Distillation, which distills the knowledge from well-trained and highly complex teacher model into a compact student model, has been widely used for model compression. However, under the strict requirement on the resource cost, it is quite challenging to achieve comparable performance with the teacher model, essentially due to the drastically-reduced expressiveness ability of the compact student model. Inspired by the nature of the expressiveness ability in Neural Networks, we propose to use multi-segment activation, which can significantly improve the expressiveness ability with very little cost, in the compact student model. Specifically, we propose a highly efficient multi-segment activation, called Light Multi-segment Activation (LMA), which can rapidly produce multiple linear regions with very few parameters by leveraging the statistical information. With using LMA, the compact student model is capable of achieving much better performance effectively and efficiently, than the ReLU-equipped one with same model scale. Furthermore, the proposed method is compatible with other model compression techniques, such as quantization, which means they can be used jointly for better compression performance. Experiments on state-of-the-art NN architectures over the real-world tasks demonstrate the effectiveness and extensibility of the LMA. |
Tasks | Model Compression, Quantization |
Published | 2019-07-16 |
URL | https://arxiv.org/abs/1907.06870v2 |
https://arxiv.org/pdf/1907.06870v2.pdf | |
PWC | https://paperswithcode.com/paper/light-multi-segment-activation-for-model |
Repo | https://github.com/LMA-NeurIPS19/LMA |
Framework | pytorch |
Learning-Driven Exploration for Reinforcement Learning
Title | Learning-Driven Exploration for Reinforcement Learning |
Authors | Muhammad Usama, Dong Eui Chang |
Abstract | Deep reinforcement learning algorithms have been shown to learn complex skills using only high-dimensional observations and scalar reward. Effective and intelligent exploration still remains an unresolved problem for reinforcement learning. Most contemporary reinforcement learning relies on simple heuristic strategies such as $\epsilon$-greedy exploration or adding Gaussian noise to actions. These heuristics, however, are unable to intelligently distinguish the well explored and the unexplored regions of the state space, which can lead to inefficient use of training time. We introduce entropy-based exploration (EBE) that enables an agent to explore efficiently the unexplored regions of the state space. EBE quantifies the agent’s learning in a state using merely state dependent action values and adaptively explores the state space, i.e. more exploration for the unexplored region of the state space. We perform experiments on many environments including a simple linear environment, a simpler version of the breakout game and multiple first-person shooter (FPS) games of VizDoom platform. We demonstrate that EBE enables efficient exploration that ultimately results in faster learning without having to tune hyperparameters. |
Tasks | Efficient Exploration, FPS Games |
Published | 2019-06-17 |
URL | https://arxiv.org/abs/1906.06890v1 |
https://arxiv.org/pdf/1906.06890v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-driven-exploration-for-reinforcement |
Repo | https://github.com/Usama1002/EBE-Exploration |
Framework | tf |
Cross-Lingual Alignment of Contextual Word Embeddings, with Applications to Zero-shot Dependency Parsing
Title | Cross-Lingual Alignment of Contextual Word Embeddings, with Applications to Zero-shot Dependency Parsing |
Authors | Tal Schuster, Ori Ram, Regina Barzilay, Amir Globerson |
Abstract | We introduce a novel method for multilingual transfer that utilizes deep contextual embeddings, pretrained in an unsupervised fashion. While contextual embeddings have been shown to yield richer representations of meaning compared to their static counterparts, aligning them poses a challenge due to their dynamic nature. To this end, we construct context-independent variants of the original monolingual spaces and utilize their mapping to derive an alignment for the context-dependent spaces. This mapping readily supports processing of a target language, improving transfer by context-aware embeddings. Our experimental results demonstrate the effectiveness of this approach for zero-shot and few-shot learning of dependency parsing. Specifically, our method consistently outperforms the previous state-of-the-art on 6 tested languages, yielding an improvement of 6.8 LAS points on average. |
Tasks | Dependency Parsing, Few-Shot Learning, Word Embeddings |
Published | 2019-02-25 |
URL | http://arxiv.org/abs/1902.09492v2 |
http://arxiv.org/pdf/1902.09492v2.pdf | |
PWC | https://paperswithcode.com/paper/cross-lingual-alignment-of-contextual-word |
Repo | https://github.com/TalSchuster/CrossLingualELMo |
Framework | pytorch |
Variational Representation Learning for Vehicle Re-Identification
Title | Variational Representation Learning for Vehicle Re-Identification |
Authors | Saghir Ahmed Saghir Alfasly, Yongjian Hu, Tiancai Liang, Xiaofeng Jin, Qingli Zhao, Beibei Liu |
Abstract | Vehicle Re-identification is attracting more and more attention in recent years. One of the most challenging problems is to learn an efficient representation for a vehicle from its multi-viewpoint images. Existing methods tend to derive features of dimensions ranging from thousands to tens of thousands. In this work we proposed a deep learning based framework that can lead to an efficient representation of vehicles. While the dimension of the learned features can be as low as 256, experiments on different datasets show that the Top-1 and Top-5 retrieval accuracies exceed multiple state-of-the-art methods. The key to our framework is two-fold. Firstly, variational feature learning is employed to generate variational features which are more discriminating. Secondly, long short-term memory (LSTM) is used to learn the relationship among different viewpoints of a vehicle. The LSTM also plays as an encoder to downsize the features. |
Tasks | Representation Learning, Vehicle Re-Identification |
Published | 2019-05-07 |
URL | https://arxiv.org/abs/1905.02343v1 |
https://arxiv.org/pdf/1905.02343v1.pdf | |
PWC | https://paperswithcode.com/paper/variational-representation-learning-for |
Repo | https://github.com/saghiralfasly/VFL-Vehicle-Re-Id |
Framework | tf |
SensitiveNets: Learning Agnostic Representations with Application to Face Recognition
Title | SensitiveNets: Learning Agnostic Representations with Application to Face Recognition |
Authors | Aythami Morales, Julian Fierrez, Ruben Vera-Rodriguez |
Abstract | This work proposes a new neural network feature representation that help to leave out sensitive information in the decision-making process of pattern recognition and machine learning algorithms. The aim of this work is to develop a learning method capable to remove certain information from the feature space without drop of performance in a recognition task based on that feature space. Our work is in part motivated by the new international regulation for personal data protection, which forces data controllers to avoid discriminative hazards while managing sensitive data of users. Our method is based on a triplet loss learning generalization that introduces a sensitive information removal process. The method is evaluated on face recognition technologies using state-of-the-art algorithms and publicly available benchmarks. In addition, we present a new annotation dataset with balanced distribution between genders and ethnic origins. The dataset includes more than 120K images from 24K identities with variety of poses, image quality, facial expressions, and illumination. The experiments demonstrate that it is possible to reduce sensitive information such as gender or ethnicity in the feature representation while retaining competitive performance in a face recognition task. |
Tasks | Face Recognition, Representation Learning |
Published | 2019-02-01 |
URL | http://arxiv.org/abs/1902.00334v1 |
http://arxiv.org/pdf/1902.00334v1.pdf | |
PWC | https://paperswithcode.com/paper/sensitivenets-learning-agnostic |
Repo | https://github.com/BiDAlab/DiveFace |
Framework | none |